Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[XLA] HloEvaluator - use TF_ASSIGN_OR_RETURN for HandleCall, HandleFusion and HandleConditional #20967

Merged
13 changes: 5 additions & 8 deletions tensorflow/compiler/xla/service/hlo_evaluator.cc
Expand Up @@ -1216,8 +1216,7 @@ Status HloEvaluator::HandleCall(HloInstruction* call) {
HloEvaluator embedded_evaluator;
embedded_evaluator.set_dynamic_dimension_inference(
dynamic_dimension_inference_);
Literal result = embedded_evaluator.Evaluate(*computation, arg_literals)
.ConsumeValueOrDie();
TF_ASSIGN_OR_RETURN(Literal result, embedded_evaluator.Evaluate(*computation, arg_literals));

evaluated_[call] = std::move(result);
return Status::OK();
Expand Down Expand Up @@ -1251,9 +1250,8 @@ Status HloEvaluator::HandleFusion(HloInstruction* fusion) {
HloEvaluator embedded_evaluator;
embedded_evaluator.set_dynamic_dimension_inference(
dynamic_dimension_inference_);
Literal result =
embedded_evaluator.Evaluate(*readded_computation, arg_literals)
.ConsumeValueOrDie();
TF_ASSIGN_OR_RETURN(Literal result,
embedded_evaluator.Evaluate(*readded_computation, arg_literals));

evaluated_[fusion] = std::move(result);
return Status::OK();
Expand All @@ -1277,10 +1275,9 @@ Status HloEvaluator::HandleConditional(HloInstruction* conditional) {
HloEvaluator embedded_evaluator;
embedded_evaluator.set_dynamic_dimension_inference(
dynamic_dimension_inference_);
Literal result = embedded_evaluator
TF_ASSIGN_OR_RETURN(Literal result, embedded_evaluator
.Evaluate(*conditional->branch_computation(branch_index),
{&branch_computation_arg})
.ConsumeValueOrDie();
{&branch_computation_arg}));

evaluated_[conditional] = std::move(result);
return Status::OK();
Expand Down