Good programs are written by people who anticipate what might go wrong. If the document says 'don't do X'; they know a tester is likely to try X because a user will eventually do it.
I can see an LLM producing a good program with terrible code that's hard to grok and adjust.