zlacker

[parent] [thread] 2 comments
1. buggle+(OP)[view] [source] 2025-01-21 23:42:06
> AI is DOA. LLMs have no successor, and the transformer architecture hit it's bathtub curve years ago

Tell me you didn’t read the DeepSeek R1 paper without telling me you also don’t know about reinforcement learning.

replies(1): >>tallda+x
2. tallda+x[view] [source] 2025-01-21 23:45:22
>>buggle+(OP)
R1 is a rehash of things we've already seen, and a particularly neutered one at that. Are there any better examples you can think of?
replies(1): >>buggle+d1
◧◩
3. buggle+d1[view] [source] [discussion] 2025-01-21 23:49:25
>>tallda+x
Uh, they invented multilatent attention and since the method for creating o1 was never published, they’re the only documented example of producing a model of comparable quality. They also demonstrated massive gains to the performance of smaller models through distillation of this model/these methods, so no, not really. I know this is the internet, but we should try to not just say things.
[go to top]