zlacker

[parent] [thread] 7 comments
1. swalsh+(OP)[view] [source] 2023-11-17 21:27:35
This is plausible to me, there's no way anyone is making money from my $20 subscription I use ALL DAY LONG.
replies(2): >>coffee+M1 >>jstumm+L3
2. coffee+M1[view] [source] 2023-11-17 21:34:43
>>swalsh+(OP)
Is inference really that expensive? Anyway if the price is too low they could easily charge by query
replies(1): >>knicho+29
3. jstumm+L3[view] [source] 2023-11-17 21:42:34
>>swalsh+(OP)
Given the arbitrary rate limiting they take the liberty of doing, it's a very deliberate decision and entirely within their control to change at any point.
◧◩
4. knicho+29[view] [source] [discussion] 2023-11-17 22:13:03
>>coffee+M1
When I was mining with a bunch of RTX 3080s and RTX 3090s, the electricity cost (admittedly) was about $20/month per card. Running a 70B model takes 3-4 cards. Assuming you're pushing these cards to their extreme max, it's going to be $80/mo. Then again, ChatGPT is pretty awesome, and is likely running more than a 70B model (or I think I heard it was running an ensemble of models), so there's at least a ballpark.
replies(3): >>sodali+3c >>Sebb76+ai >>698969+Qa1
◧◩◪
5. sodali+3c[view] [source] [discussion] 2023-11-17 22:27:59
>>knicho+29
Batched inference makes these calculations hard - roughly takes the same amount of power and time for one inference vs 30 (as i understand it)
◧◩◪
6. Sebb76+ai[view] [source] [discussion] 2023-11-17 22:54:13
>>knicho+29
Datacenters probably do not pay retail rates on electricity, so they might actually run quite a bit cheaper (or more expensive if they use highly available power, but this seems like overkill for pure compute power).
replies(1): >>015a+HH
◧◩◪◨
7. 015a+HH[view] [source] [discussion] 2023-11-18 00:57:52
>>Sebb76+ai
Sure, but everything else about a data center is more expensive (real estate, operations people, networking, equipment). There's a reason AWS is so expensive.
◧◩◪
8. 698969+Qa1[view] [source] [discussion] 2023-11-18 04:24:22
>>knicho+29
Presumably your miner is running 24/7 throughout the month. Not the same for ChatGPT which would answer maybe 10 sessions (with multiple pauses between queries) tops from a single person in a day.
[go to top]