Skip to content

Commit

Permalink
Add max_enqueued_batches option for model servers
Browse files Browse the repository at this point in the history
PiperOrigin-RevId: 654998187
  • Loading branch information
deqiangc authored and tensorflow-copybara committed Jul 23, 2024
1 parent 6b9cf7c commit 7c99259
Show file tree
Hide file tree
Showing 2 changed files with 5 additions and 0 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -236,6 +236,8 @@ absl::Status TfrtSavedModelFactory::CreateTfrtSavedModelWithMetadata(
compile_options.use_gpu_compile_and_execute_op =
config_.tfrt_use_fused_gpu_op();
compile_options.min_num_batch_threads = config_.tfrt_min_num_batch_threads();
compile_options.min_max_enqueued_batches =
config_.tfrt_min_max_enqueued_batches();

options.graph_execution_options.run_placer_grappler_on_functions =
config_.run_placer_grappler_on_functions();
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -221,6 +221,9 @@ message TfrtSavedModelConfig {
IFRT_CLIENT_PATHWAYS = 2;
}
IfrtClientType ifrt_client_type = 2024;

// The minimum of the maximum number of outstanding enqueue batches
int64 tfrt_min_max_enqueued_batches = 2025;
}

// Config proto for TfrtSavedModelSourceAdapter.
Expand Down

0 comments on commit 7c99259

Please sign in to comment.