For about 1000 input tokens (and resulting 1000 output tokens), to my surprise, GPT-3.5 turbo was 100x cheaper than Llama 2.
Llama 7B wasn't up to the task fyi, producing very poor translations.
I believe that OpenAI priced GPT-3.5 aggressively cheap in order to make it a non-brainer to rely on them rather than relying on other vendors (even open source models).
I'm curious to see if others have gotten different results?
But you are totally correct about the pricing part it can get expensive
I’m running this photo service https://msdosimagetools.ngrok.dev/
Its doing 200+ photos every day and I’m using open source models behind the scene on replicate. My costs increasing day by day
Plus this is hosted locally