zlacker

[parent] [thread] 1 comments
1. kadush+(OP)[view] [source] 2025-04-05 16:16:54
There are LLMs which do not generate one token at a time: https://arxiv.org/abs/2502.09992

They do not reason significantly better than autoregressive LLMs. Which makes me question “one token at a time” as the bottleneck.

Also, Lecun has been pushing his JEPA idea for years now - with not much to show for it. With his resources one could hope we would see the benefits of that over the current state of the art models.

replies(1): >>financ+88
2. financ+88[view] [source] 2025-04-05 17:26:44
>>kadush+(OP)
from the article: LeCun has been working in some way on V-JEPA for two decades. At least it's bold, and, everyone says it won't work until one day it might
[go to top]