← Back to context

Comment by simonw

25 days ago

Because I've played this game too many times before - I know that some people will find a hole in any example you show them.

So before doing that work, I want to get a feel for if you're asking this question in good faith and have done any active looking yourself.

(My favorite two recent open source examples are https://simonwillison.net/2026/Jan/27/one-human-one-agent-on... and https://github.com/antirez/flux2.c)

>>What's the open source codebase generated entirely by AI that has impressed you the most?

>One Human + One Agent = One Browser From Scratch

I at least expect you to read my post before replying.

  • What do you mean? Are you suggesting that the "one human" means it wasn't entirely written by AI?

    That's not the case, the "one human" there is the one human prompting it: https://emsh.cat/one-human-one-agent-one-browser/

    If your goalpost here is "no human involved at all" then it's a good thing I asked you what your goalposts were before spending any time on this!

    UPDATE: OK I think I see what's happened here! You're asking to see an open source repo that was built using the "dark factory" pattern, where no code was even reviewed by a human.

    I don't think I've seen one of those yet - I mean maybe that Cursor FastRender thing comes close?

    It's a very radical technique. I don't think many people are trying this yet - I haven't been brave enough to try it myself yet.

    I guess I kind of did that with my Python WASM library? That was an experiment in how far I could get with prompting and not reviewing, but it's not something I'd hold up as a shining example of how projects should be built: https://github.com/simonw/pwasm

    • Your original post in the thread is about an automated dark factory with thousands of AI agents. It's amazing but we can't see it because they are in stealth mode.

      Then the first example of a project done by AI without human intervention is someone who _explicitly_ states that they drove the way the agent behaved.

      From the blog:

      >The human who drives the agent might matter more than how the agents work and are set up, the judge is still out on this one

      >If one person with one agent can produce equal or better results than "hundreds of agents for weeks", then the answer to the question: "Can we scale autonomous coding by throwing more agents at a problem?", probably has a more pessimistic answer than some expected.

      I'm really not understanding what this proves other than the fact that AI + human is great and AI + AI is shit. Something that both me and the person who did the browser agreed on: https://news.ycombinator.com/item?id=46783282

      3 replies →