You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
On 1. prompt_token_counts would be the same behavior as per-request max_output_len?
On 2> logprobs is a boolean arg on input to signify to return the logprobs in the response protocol buffer?
On 3> This would be a global seed as command line argument to start the server; as seed itself is a global in torch?
Currently sampling params such as temperature are set as commandline flags in when the server starts.
It would be nice for each request to pass in the sampling params instead.
The text was updated successfully, but these errors were encountered: