zlacker

[return to "Scaling long-running autonomous coding"]
1. embedd+Ya[view] [source] 2026-01-14 23:02:00
>>samwil+(OP)
Did anyone manage to run the tests from the repository itself? The code seems filled with errors and warnings, as far as I can tell none of them because of the platform I'm on (Linux). I went and looked at the Action workflow history for some pages, and seems CI been failing for a while, PRs also all been failing CI but merged. How exactly was this verified to be something to be used as an successful example, or am I misunderstanding what point they are trying to make? They mention a screenshot, but they never actually mention if their goal was successfully met, do they?

I'm not sure the approach of "completely autonomous coding" is the right way to go. I feel like maybe we'll be able to use it more effectively if we think of them as something to be used by a human to accomplish some thing instead, lean into letting the human drive the thing instead, because quality spirals so quickly out of control.

◧◩
2. csomar+271[view] [source] 2026-01-15 06:30:28
>>embedd+Ya
You can stop reading the article from here:

> Today's agents work well for focused tasks, but are slow for complex projects.

What does slow mean? Slower than humans? Need faster GPUs? What does it even imply? Too slow to produce the next token? Too slow in attempts to be usable? Need human intervention?

This piece is made and written to keep the bubble inflating further.

[go to top]