Unfortunately, if a single inference is unable to complete in the allotted API call interval, you will see timeouts. I believe most inferences created using Coda’s AI is a single request, so you’re really at the mercy of the LLM.
I grew tired of AI systems that were sluggish and switched to Google’s PaLM 2. It seems to be about 3X faster responding and I have seen only a few times when it was busy enough to result in a failed request.
Since it also requires no paid account, I tend to break my AI processes into multiple stages. This speeds up each request while also giving me the latitude to perform long-running automations. I have a few agents that will run for 10 minutes and not time out. [more on that topic here]
While OpenAI is the leader in LLM capabilities, I have learned that many of the open-source LLMs are powerful enough in many use cases. Ideally, Coda will make it’s AI features LLM-agnostic allowing us to choose from the dozens (soon to be hundreds) of inferencing options avialable.