Comment by the-lazy-guy
19 days ago
That is exactly what you want to evaluate the thechnology. Not make a buggy commit into softwared not used by nobody and reviewed by an intern. But actually review it by domain professionals, in real world very well-tested project. So they could make an informed decision on where it lacks in capabilities and what needs to be fixed before they try it again.
I doubt that anyone expected to merge any of these PRs. Question is - can the machine solve minor (but non-trivial) issues listed on github in an efficient way with minimal guidance. Current answer is no.
Also, _if_ anything was to be merged, dotnet is dogfooded extensively at Microsoft, so bugs in it are much more likely to be noticed and fixed before you get a stable release on your plate.
> Not make a buggy commit into software not used by nobody and reviewed by an intern.
If it can't even make a decent commit into software nobody uses, how can it ever do it for something even more complex? And no, you don't need to review it with an intern...
> can the machine solve minor (but non-trivial) issues listed on github in an efficient way with minimal guidance
I'm sorry but the only way this is even a question is if you never used AI in the real world. Anyone with a modicum of common sense would tell you immediately: it cannot.
You can't even keep it "sane" in a small conversation, let alone using tons of context to accomplish non-trivial tasks.