zlacker

[return to "Watching AI drive Microsoft employees insane"]
1. margor+72[view] [source] 2025-05-21 11:23:29
>>laiysb+(OP)
With how stochastic the process is it makes it basically unusable for any large scale task. What's the plan? To roll the dice until the answer pops up? That would be maybe viable if there was a way to automatically evaluate it 100% but with a human in the loop required it becomes untenable.
◧◩
2. eterev+J2[view] [source] 2025-05-21 11:33:07
>>margor+72
The plan is to improve AI agents from their current ~intern level to a level of a good engineer.
◧◩◪
3. ethano+l3[view] [source] 2025-05-21 11:38:38
>>eterev+J2
Seems like that is taking a very long time, on top of some very grandiose promises being delivered today.
◧◩◪◨
4. infect+K5[view] [source] 2025-05-21 11:58:42
>>ethano+l3
I look back over the past 2-3 years and am pretty amazed with how quick change and progress have been made. The promises are indeed large but the speed of progress has been fast. Not defending the promise but “taking a very long time” does not seem to be an accurate representation.
◧◩◪◨⬒
5. ethano+t6[view] [source] 2025-05-21 12:04:41
>>infect+K5
I guess it probably depends on what you are doing. Outside of layers on top of these things (tooling), I personally haven't seen much progress.
◧◩◪◨⬒⬓
6. infect+F8[view] [source] 2025-05-21 12:20:48
>>ethano+t6
What a time we live in. I guess it depends how pessimistic you are.
◧◩◪◨⬒⬓⬔
7. lcnPyl+Wb[view] [source] 2025-05-21 12:45:59
>>infect+F8
To their point, there hasn’t been any huge breakthrough in this field since the “attention is all you need” paper. Not really any major improvements to model architecture, as far as I am aware. (Admittedly, this is a new field of study to me.) I believe one hope is to develop better methods for self-supervised learning; I am not sure of the progress there. Most practical improvements have been on the hardware and tooling side (GPUs and, e.g., pytorch).

Don’t get me wrong: the current models are already powerful and useful. However, there is still a lot of reason to remain skeptical of an imminent explosion in intelligence from these models.

◧◩◪◨⬒⬓⬔⧯
8. infect+Uc[view] [source] 2025-05-21 12:53:23
>>lcnPyl+Wb
You’re totally right that there hasn’t been a fundamental architectural leap like “attention is all you need”, that was a generational shift. But I’d argue that what we’ve seen since is a compounding of scale, optimization, and integration that’s changed the practical capabilities quite dramatically, even if it doesn’t look flashy in an academic sense. The models are qualitatively different at the frontier, more steerable, more multimodal, and increasingly able to reason across context. It might not feel like a revolution on paper, but the impact in real-world workflows is adding up quickly. Perhaps all of that can be put in the bucket of “tooling” but from my perspective there has still been quite large leaps looking at cost differences alone.

For some reason my pessimism meter goes off when I see single sentence arguments “change has been slow”. Thanks for brining the conversation back.

[go to top]