zlacker
[parent]
[thread]
2 comments
1. Comput+(OP)
[view]
[source]
2023-12-20 23:53:10
But this doesn’t apply to self-hosted, o?
replies(1):
>>lyjack+Le
◧
2. lyjack+Le
[view]
[source]
2023-12-21 01:54:39
>>Comput+(OP)
It does. LLMs are most efficient when running large batches, so the gpu cost is super high if you’re underutilizing it. It will cost more than a cloud provider like open ai who has the volume to keep their GPUs saturated
replies(1):
>>jay-ba+jt
◧◩
3. jay-ba+jt
[view]
[source]
[discussion]
2023-12-21 04:35:54
>>lyjack+Le
Yup. It’s also important to mention that OpenAI enjoys the luxury of having large clusters of H100s (the last time I checked).
[go to top]