zlacker

[return to "Qwen3-Coder-Next"]
1. skhame+l9[view] [source] 2026-02-03 16:38:51
>>daniel+(OP)
It’s hard to elaborate just how wild this model might be if it performs as claimed. The claims are this can perform close to Sonnet 4.5 for assisted coding (SWE bench) while using only 3B active parameters. This is obscenely small for the claimed performance.
◧◩
2. cirrus+ab[view] [source] 2026-02-03 16:45:25
>>skhame+l9
If it sounds too good to be true…
◧◩◪
3. theshr+af[view] [source] 2026-02-03 17:01:33
>>cirrus+ab
Should be possible with optimised models, just drop all "generic" stuff and focus on coding performance.

There's no reason for a coding model to contain all of ao3 and wikipedia =)

◧◩◪◨
4. moffka+zz[view] [source] 2026-02-03 18:22:15
>>theshr+af
That's what Meta thought initially too, training codellama and chat llama separately, and then they realized they're idiots and that adding the other half of data vastly improves both models. As long as it's quality data, more of it doesn't do harm.

Besides, programming is far from just knowing how to autocomplete syntax, you need a model that's proficient in the fields that the automation is placed in, otherwise they'll be no help in actually automating it.

[go to top]