-
Notifications
You must be signed in to change notification settings - Fork 751
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Rate Limit being hit consistently #123
Comments
Parsing the rate limit error and waiting accordingly is a good idea. I'll look into it. |
Worth noting that after you have spent $50 total on your account, the limit increases to 450k tpm. |
Same here. How to tell amount of usage by model in the set? Checked OpenAI usage and nothing alarming there. |
if you are running it locally, you can try changing the retry waitBackoff func to apply 5 second additional for each numRetry. I got a lot less of them that way. For heavier things, might could use 10 sec. Lastly, this might be a great use case for balancing between anthropic and OpenAI “agents”. |
@atljoseph I decreased the backoff a bit in the last release so I may need to revert that or make it configurable. Or just parse the error message and wait accordingly as @cfieandres suggested. My token limit is quite high from building/testing Plandex so I'm not getting any of these errors. It’s helpful to know what backoff is working for you at a lower limit--thanks. |
It’s working great. I’ll take a bit slower any day instead of 429s. Yeah we
do some parsing similar to that at work regexp is great for this use case.
…On Sun, May 19, 2024 at 1:01 PM Dane Schneider ***@***.***> wrote:
@atljoseph <https://github.com/atljoseph> I decreased the backoff a bit
in the last release so I may need to revert that or make it configurable.
Or just parse the error message and wait accordingly as @cfieandres
<https://github.com/cfieandres> suggested. My token limit is quite high
from building/testing Plandex so I'm not getting any of these errors, so
it's helpful to know what backoff is working for you at a lower
limit--thanks.
—
Reply to this email directly, view it on GitHub
<#123 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AF632FOMUE6GMQ73LKCX4YLZDDLGZAVCNFSM6AAAAABH4343WWVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDCMJZGMYDANRWHA>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
Well, it does sometimes hit this error. After a few retried 429 logs in server is when i noticed it... immediately after the subsequent successful retry... No idea if these events are connected or not. Getting The, it leads to |
As of |
@danenania Would you mind increasing the limit to 60 seconds? |
After updating to 1.0.0, plandex has consistently hit openai's rate limit, which makes it so that I need to do "plandex c" frequently.
Is there a way to have plandex wait till not rate limited and continue by itself?
🚨 Server error
→ Error starting reply stream
→ Error, status code
→ 429, message
→ Rate limit reached for gpt-4o in organization xxxxxxxxxx on tokens per min (TPM)
→ Limit 30000, Used 23193, Requested 15447. Please try again in 17.28s. Visit https://platform.openai.com/account/rate-limits to learn more.
The text was updated successfully, but these errors were encountered: