We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
CreateChatCompletionRequest.max_tokens is a Option<u16> as of 0.23.1.
CreateChatCompletionRequest.max_tokens
Option<u16>
The newer models such as gpt-4o has a context window of 128,000 tokens. This context window limit is the sum of input and output tokens.
I believe the max_tokens field should be Option<u32> to allow numbers as high as 128,000.
max_tokens
Option<u32>
The text was updated successfully, but these errors were encountered:
Thank you for reporting the bug, a PR is most welcome!
Sorry, something went wrong.
Successfully merging a pull request may close this issue.
CreateChatCompletionRequest.max_tokens
is aOption<u16>
as of 0.23.1.The newer models such as gpt-4o has a context window of 128,000 tokens. This context window limit is the sum of input and output tokens.
I believe the
max_tokens
field should beOption<u32>
to allow numbers as high as 128,000.The text was updated successfully, but these errors were encountered: