zlacker

[return to "AI tooling must be disclosed for contributions"]
1. Waterl+A3[view] [source] 2025-08-21 19:07:52
>>freeto+(OP)
I’m not a big AI fan but I do see it as just another tool in your toolbox. I wouldn’t really care how someone got to the end result that is a PR.

But I also think that if a maintainer asks you to jump before submitting a PR, you politely ask, “how high?”

◧◩
2. cvoss+56[view] [source] 2025-08-21 19:24:50
>>Waterl+A3
It does matter how and where a PR comes from, because reviewers are fallible and finite, so trust enters the equation inevitably. You must ask "Do I trust where this came from?" And to answer that, you need to know where it come from.

If trust didn't matter, there wouldn't have been a need for the Linux Kernel team to ban the University of Minnesota for attempting to intentionally smuggle bugs through the PR process as part of an unauthorized social experiment. As it stands, if you / your PRs can't be trusted, they should not even be admitted to the review process.

◧◩◪
3. otterl+0B[view] [source] 2025-08-21 22:27:54
>>cvoss+56
If it comes with good documentation and appropriate tests, does that help?
◧◩◪◨
4. mattbe+kH[view] [source] 2025-08-21 23:08:27
>>otterl+0B
The observation that inspired this policy is that if you used AI, it is likely you don't know if the code, the documentation or tests are good or appropriate.
◧◩◪◨⬒
5. otterl+qI[view] [source] 2025-08-21 23:17:41
>>mattbe+kH
What if you started with good documentation that you personally wrote, you gave that to the agent, and you verified the tests were appropriate and passed?
[go to top]