zlacker

[parent] [thread] 7 comments
1. ramesh+(OP)[view] [source] 2023-09-12 19:27:18
>For about 1000 input tokens (and resulting 1000 output tokens), to my surprise, GPT-3.5 turbo was 100x cheaper than Llama 2.

You'll never get actual economics out of switching to open models without running your own hardware. That's the whole point. There's orders of magnitude difference in price, where a single V100/3090 instance can run llama2-70b inference for ~0.50$/hr.

replies(1): >>YetAno+J5
2. YetAno+J5[view] [source] 2023-09-12 19:46:33
>>ramesh+(OP)
No, they can't run it. llama 70 with 4 bit quantization takes ~50 GB VRAM for decent enough context size. You need A100, or 2-3 V100 or 4 3090 which all costs roughly roughly $3-5/h
replies(1): >>ramesh+a7
◧◩
3. ramesh+a7[view] [source] [discussion] 2023-09-12 19:50:28
>>YetAno+J5
Wrong. I am running 8bit GGML with 24GB VRAM on a single 4090 with 2048 context right now
replies(1): >>YetAno+x7
◧◩◪
4. YetAno+x7[view] [source] [discussion] 2023-09-12 19:51:47
>>ramesh+a7
Which model? I am talking about 70b as mentioned clearly. 70b 8b is 70GB just for the model itself. How much token/second are you getting with single 4090?
replies(1): >>ramesh+G8
◧◩◪◨
5. ramesh+G8[view] [source] [discussion] 2023-09-12 19:55:38
>>YetAno+x7
Offloading 40% of layers to CPU, about 50t/s with 16 threads.
replies(2): >>pocket+nh >>jpdus+La1
◧◩◪◨⬒
6. pocket+nh[view] [source] [discussion] 2023-09-12 20:24:37
>>ramesh+G8
That is more than an order of magnitude better than my experience; I get around 2 t/s with similar hardware. I had also seen others reporting similar figures to mine so I assumed it was normal. Is there a secret to what you're doing?
replies(1): >>ramesh+sM
◧◩◪◨⬒⬓
7. ramesh+sM[view] [source] [discussion] 2023-09-12 22:42:09
>>pocket+nh
>Is there a secret to what you're doing?

Core speed and memory bandwidth matter a lot. This is on a Ryzen 7950 with DDR5.

◧◩◪◨⬒
8. jpdus+La1[view] [source] [discussion] 2023-09-13 01:36:16
>>ramesh+G8
Care to share your detailed stack and command to reach 50t/s? I also have a 7950 with DDR 5 and I don't even get 50 t/s on my two RTX 4090s....
[go to top]