This post was split from the topic:
I’m finding that if I use a short prompt, limit it to 1000 tokens output on the openAI gpt-4o model, with a new assistant each time I am receiving the error
Failed to create a conversation in ChatGPT
Error from halted execution: Request too large for gpt-4o in organization org-xxxx on tokens per min (TPM): Limit 30000, Requested 41839.
With the requested amount increasing on each test.
I have credit, and my request runs fine on the openAI site.
Is there some cache or something?