zlacker

[parent] [thread] 1 comments
1. padols+(OP)[view] [source] 2023-11-18 08:31:51
If you're just using their completions/chat API, you're gonna be ok. As an ultimate fallback you can spin up H100s in the cloud and run VLLM atop a high param open model like Llama 70B. Such models will catch up and their param counts will increase.. eventually. But initially expect gpt-3.5-esque performance. VLLM will give you an OpenAI-like REST API atop a range of models. Keep making things :))
replies(1): >>karmas+n
2. karmas+n[view] [source] 2023-11-18 08:34:54
>>padols+(OP)
Thx. I will. My current interests mainly lies in benchmarking their vision model.

That being said, I might not go further relying on their APIs for something more serious

[go to top]