zlacker

[parent] [thread] 2 comments
1. Comput+(OP)[view] [source] 2023-12-20 23:53:10
But this doesn’t apply to self-hosted, o?
replies(1): >>lyjack+Le
2. lyjack+Le[view] [source] 2023-12-21 01:54:39
>>Comput+(OP)
It does. LLMs are most efficient when running large batches, so the gpu cost is super high if you’re underutilizing it. It will cost more than a cloud provider like open ai who has the volume to keep their GPUs saturated
replies(1): >>jay-ba+jt
◧◩
3. jay-ba+jt[view] [source] [discussion] 2023-12-21 04:35:54
>>lyjack+Le
Yup. It’s also important to mention that OpenAI enjoys the luxury of having large clusters of H100s (the last time I checked).
[go to top]