Replies: 1 comment
-
|
The max_tokens parameter should work with AsyncOpenAI. Can you check a few things: First, verify it's actually limiting the response. Try something like this: If max_tokens is working, you should see:
A couple things that might be happening:
What model are you using and what token count are you seeing in the response? |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I would like to bound the maximum number of tokens of the response to help with formatting when working with AsyncOpenAI, but I could not find anything about it.
I tried :
but the max_tokens argument is seemingly ignored.
Beta Was this translation helpful? Give feedback.
All reactions