Love Coda Ai in so many ways and I have one important piece of feedback:
The native Coda Ai function is SOOO much easier to use overall than the Open AI pack.
BUT I’m using the Open AI pack way more because of the models. Which is kind of frustrating because it’s so much more difficult to use.
The Coda Ai models seem to be some old complete model like davinci or something, or a model that has been fine tuned for Coda. I totally get this decision, especially from a cost perspective on Coda’s side. Right now it’s still all free.
But the model is so bad compared to the models I want to actually use (3.5 16k and soon to be GPT-4), that I just don’t use it for most things.
I would be AMAZING if I could just plug my API key into the native coda model and use the model I want and pay for it myself. I realize that maybe this isn’t so possible with how it’s build, but just FYI.
This gap between usability will grow even more extreme when GPT-4 API comes out in a week. And it just makes me sad to have to still use the pack all the time, when I now how awesome the native AI is.
For example I explicitly told the native AI to NOT use “protein” in a response or ANY numbers. Its response:
“A medium-sized serving of chicken breast with a protein content that is comparable to a 5 oz steak.”
No matter what I seem to say it just can’t do it, which just makes it not usable for me. And there so so many of these examples. Whereas GPT 3.5 out of the pack runs absolute laps around the Coda model, and gets the job done so much better.
interesting feedback
I created my own google AI pack and the payment runs via my own keys in the google console
it works, but since it is not native, it is not so useful.
as @Bill_French mentioned elsewhere it would be great if we could select our the AI of our choice and I added that I’d like to see options like below included we can play with per model, per usage.
But for whatever reason it vastly different results, so it might GPT3 or just a trained in a particular way. Coda ai seems to do a better job of responding to staying within sentence lengths, and some other things. But then there are outputs that are just very bad. Maybe the temperature is set a bit lower than normal too. Not sure but it doesn’t seem so great with a lot of things.