zlacker

[parent] [thread] 2 comments
1. IhateA+(OP)[view] [source] 2026-02-03 17:32:47
Imagine thinking that $80k setups to run Kimi and serve a single user session is evidence that inference providers are running at cost, or even close to it. Or that this fact is some sort of proof that token pricing will come down. All you one-shotted llm dependents said the same thing about Deepseek.

I know you need to cope because your competency is 1:1 correlated to the quality and quantity of tokens you can afford, so have fun with your Think for me SaaS while you can afford it. You have no clue the amount of engineering that goes into provide inference at scale. I wasn't even including the cost of labor.

replies(1): >>deaux+dJ1
2. deaux+dJ1[view] [source] 2026-02-04 02:37:41
>>IhateA+(OP)
It directly disproves this wild claim

> You still need $500k in GPUs and a boatload of electricity to serve like 3 concurrent sessions at a decent tok/ps.

as being patent bullshit, after which the burden is squarely on you to back up the remainder of your claims.

replies(1): >>IhateA+fP1
◧◩
3. IhateA+fP1[view] [source] [discussion] 2026-02-04 03:31:06
>>deaux+dJ1
You are literally telling me that an open source model costs $80k "at decent tok/ps (whatever that means)" to run a single session as proof something. How come people aren't dropping Anthropic for Kimi, it costs 10x less... You aren't a serious person worth engaging with.
[go to top]