zlacker

[return to "My AI skeptic friends are all nuts"]
1. a_bono+vq[view] [source] 2025-06-02 23:58:23
>>tablet+(OP)
I find the Konwinski Prize to be very interesting in this context. 1 million dollars to whoever's open source LLM solves >90% of a set of novel Github issues.

https://www.kaggle.com/competitions/konwinski-prize/

Currently, the #1 spot sits at a score of 0.09, not 0.9. A far cry from being useful. I know that open source models are not as good as closed source, but still, we're a long way from LLMs being good for code on their own.

And that supports OP's point - these tools aren't AGI, they produce trash that needs evaluation, but they're still useful.

◧◩
2. naaski+vG[view] [source] 2025-06-03 02:31:36
>>a_bono+vq
> Currently, the #1 spot sits at a score of 0.09, not 0.9. A far cry from being useful.

The best intellisense and code completion tools would solve 0.00. Those were the only tools we were using just a couple of years ago. 0.09 is a tremendous jump and the improvements will accelerate!

◧◩◪
3. kortil+3I[view] [source] 2025-06-03 02:47:34
>>naaski+vG
Assuming acceleration or even continued improvement is pretty naive.
◧◩◪◨
4. naaski+Y42[view] [source] 2025-06-03 15:23:49
>>kortil+3I
You mean assuming technology improves, which it has for centuries and given there is considerable incentive for it to improve, is naive? Seems inevitable.

Do you think humans have achieved peak intelligence? If so, why, and if not, then why shouldn't you expect artificial forms of intelligence to improve up to and even surpass humans abilities at some point?

Edit: to clarify, I'm not necessarily assuming unbounded acceleration, but tools always start out middling, improvements accelerate as we figure out what works and what doesn't, and then they taper off. We're just starting on the acceleration curve for AI.

[go to top]