This is what's happening right now: they are having to review every single line produced by this machine and trying to understand why it wrote what it wrote.
Even with experienced developers reviewing and lots of tests, the likelihood of bugs in this code compared to a real engineer working on it is much higher.
Why not do this on less mission critical software at the very least?
Right now I'm very happy I don't write anything on .NET if this is what they'll use as a guinea pig for the snake oil.
I doubt that anyone expected to merge any of these PRs. Question is - can the machine solve minor (but non-trivial) issues listed on github in an efficient way with minimal guidance. Current answer is no.
Also, _if_ anything was to be merged, dotnet is dogfooded extensively at Microsoft, so bugs in it are much more likely to be noticed and fixed before you get a stable release on your plate.
If it can't even make a decent commit into software nobody uses, how can it ever do it for something even more complex? And no, you don't need to review it with an intern...
> can the machine solve minor (but non-trivial) issues listed on github in an efficient way with minimal guidance
I'm sorry but the only way this is even a question is if you never used AI in the real world. Anyone with a modicum of common sense would tell you immediately: it cannot.
You can't even keep it "sane" in a small conversation, let alone using tons of context to accomplish non-trivial tasks.