You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In this batch ingestion RFC, we proposed a batch ingestion feature which could accelerate the ingestion with neural search processors. It introduces an additional parameter "batch size" that texts from different documents could be combined and sent to ML server in one request. Since user could have different data set, different ML servers with different resources, in order to achieve better performance, they would need to experiment with different value of batch size to get the optimal performance. To offload the burden from user, we'd like to have a automation tool which could find this optimal batch size automatically.
What solution would you like?
The automation tool would run bulk index with different batch size to see which batch size would lead to optimal performance (high throughput & low latency & no errors). The OpenSearch-benchmark tool already provides rich features on benchmark which we could utilize for this automation. We can call benchmark with different parameter, collect and evaluate results then provide the recommendation.
The tool can be made to help select bulk size and client number as well which could be supported in the future phase.
What alternatives have you considered?
No alternatives.
Do you have any additional context?
No
The text was updated successfully, but these errors were encountered:
@chishui this is an interesting feature, and +1 on building such a tool. I would love to see more details around this tool to be added in the issue description(something like an RFC).
Is your feature request related to a problem?
In this batch ingestion RFC, we proposed a batch ingestion feature which could accelerate the ingestion with neural search processors. It introduces an additional parameter "batch size" that texts from different documents could be combined and sent to ML server in one request. Since user could have different data set, different ML servers with different resources, in order to achieve better performance, they would need to experiment with different value of batch size to get the optimal performance. To offload the burden from user, we'd like to have a automation tool which could find this optimal batch size automatically.
What solution would you like?
The automation tool would run bulk index with different batch size to see which batch size would lead to optimal performance (high throughput & low latency & no errors). The OpenSearch-benchmark tool already provides rich features on benchmark which we could utilize for this automation. We can call benchmark with different parameter, collect and evaluate results then provide the recommendation.
The tool can be made to help select bulk size and client number as well which could be supported in the future phase.
What alternatives have you considered?
No alternatives.
Do you have any additional context?
No
The text was updated successfully, but these errors were encountered: