Comment by azan_

22 days ago

> The pattern that gets missed in these discussions: every "no-code will replace developers" wave actually creates more developer jobs, not fewer.

Doesn't mean it will happen this time (i.e. if AI truly becomes what was promised) and actually it's not likely it will!

I felt like the article had a good argument for why the AI hype will similarly be unsuccessful at erasing developers.

> AI changes how developers work rather than eliminating the need for their judgment. The complexity remains. Someone must understand the business problem, evaluate whether the generated code solves it correctly, consider security implications, ensure it integrates properly with existing systems, and maintain it as requirements evolve.

What is your rebuttal to this argument leading to the idea that developers do need to fear for their job security?

  • No previous tool was able to learn on its own mistakes (RLVR).

    It might be not enough by itself, but it shows that something has changed in comparison with the 70-odd previous years.

    • LLM's don't learn on their own mistakes in the same way that real developers and businesses do, at least not in a way that lends itself to RLVR.

      Meaningful consequences of mistakes in software don't manifest themselves through compilation errors, but through business impacts which so far are very far outside of the scope of what an AI-assisted coding tool can comprehend.

      1 reply →

  • My argument would be that while some complexity remains, it might not require a large team of developers.

    What previously needed five devs, might be doable by just two or three.

    In the article, he says there are no shortcuts to this part of the job. That does not seem likely to be true. The research and thinking through the solution goes much faster using AI, compared to before where I had to look up everything.

    In some cases, agentic AI tools are already able to ask the questions about architecture and edge cases, and you only need to select which option you want the agent to implement.

    There are shortcuts.

    Then the question becomes how large the productivity boost will be and whether the idea that demand will just scale with productivity is realistic.

  • > evaluate whether the generated code solves it correctly, consider security implications, ensure it integrates properly with existing systems, and maintain it as requirements evolve

    I think you are basing your reasoning on the current generation of models. But if future generation will be able to do everything you've listed above, what work will be there left for developers? I'm not saying that we will ever get such models, just that when they appear, they will actually displace developers and not create more jobs for them. The business problem will be specified by business people, and even if they get it wrong it won't matter because iteration will be quick and cheap.

    > What is your rebuttal to this argument leading to the idea that developers do need to fear for their job security?

    The entire argument is based on assumption that models won't get better and will never be able to do things you've listed! But once they become capable of these things - what work will be there for developers?

> if AI truly becomes what was promised

I mean they are promising AGI.

Of course in that case it will not happen this time. However, in that case software dev getting automated would concern me less than the risk of getting turned into some manner of office supply.

Imo as long as we do NOT have AGI, software-focused professional will stay a viable career path. Someone will have to design software systems on some level of abstraction.