← Back to context

Comment by buu700

10 months ago

Understanding or being interesting has nothing to do with it. We use calculators and computers for a reason. No one hires people to respond to API requests by hand; we run the code on servers. Using the right tool for the job is just doing my job well.

> We use calculators and computers for a reason. No one hires people to respond to API requests by hand; we run the code on servers

We were talking about writing, not about vibe coding. We don't use calculators for writing. We don't use API requests for writing (except when we make an LLM write for us).

> Using the right tool for the job is just doing my job well.

I don't know what your job is. But if your job is to produce text that is meant to be read by humans, then it feels like not being able to synthesise your ideas yourself doesn't make you excellent at doing your job.

Again maybe it makes you productive. Many developers, for instance, get paid for writing bad code (either because those who pay don't care about quality or can't make a difference, or something else). Vibe coding obviously makes those developers more productive. But I don't believe it will make them learn how to produce good code. Good for them if they make money like this, of course.

  • > We were talking about writing, not about vibe coding. We don't use calculators for writing. We don't use API requests for writing (except when we make an LLM write for us).

    We do however use them to summarize and transform data all the time. Consider the ever present spreadsheet. Huge amounts of data are thrown into spreadsheets and formulas are applied to that data to present us with graphs and statistics. You could do all of that by hand, and you'd probably have a much better "internalization" about what the data is. But most of the time, hand crafting graphs from raw data and internalizing it isn't useful or necessary to accomplish what you actually want to accomplish with the data.

    • You seem to not make the difference between maths and, say, literature or history.

      Do you actually think that an LLM can take, say, a Harry Potter book as an input, and give it a grade in such a way that everybody will always agree on?

      And to go further, do you actually use LLMs to generate graphs and statistics from spreadsheet? Because that is probably a bad idea given that there are tools that actually do it right.

      2 replies →

  • We were talking about writing, not about vibe coding.

    No one said anything about vibe coding. Using tools appropriately to accomplish tasks more quickly is just common sense. Deliberately choosing to pay 10x the cost for the same or equivalent output isn't a rational business decision, regardless of whether the task happens to be writing, long division, or anything else.

    Just to be clear, I'm not arguing against doing things manually as a learning exercise or creative outlet. Sometimes the journey is the point; sometimes the destination is the point. Both are valid.

    I don't know what your job is.

    Here's one: prepping first drafts of legal docs with AI assistance before handing them off to lawyers for revision has objectively saved significant amounts of time and money. Without AI this would have been too time-consuming to be worthwhile, but with AI I've saved not only my own time but the costs of billable hours on phone calls to discuss requirements, lawyers writing first drafts on their own, and additional Q&A and revisions over email. Using AI makes it practical to skip the first two parts and cut down on the third significantly.

    Here's another one: doing security audits of customer code bases for a company that currently advertises its use of AI as a cost-saving/productivity-enhancing mechanism. Before they'd integrated AI into their platform, I would frequently get rave reviews for the quality and professionalism of my issue reports. After they added AI writing assistance, nothing changed other than my ability to generate a greater number of reports in the same number of billable hours. What you're suggesting effectively amounts to choosing to deliver less value out of ego. I still have to understand my own work product, or I wouldn't be able to produce it even with AI assistance. If someone thinks that somehow makes the product less "interesting", well then I guess it's a good thing my job isn't entertainment.

    • Don't get me wrong: I don't deny that LLMs can help tricking other humans into believing that the text is more professional than it actually is. LLMs are engineered exactly for that.

      I'd be curious to know whether your legal documents are as good as without LLMs. I wouldn't be surprised at all if they were worse, but cheaper. Talking about security audits, that's actually a problem I've seen: LLMs makes it harder to detect bad audits, and in my experience I have been more often confronted to bad security audits than to good ones.

      For both examples, you say "LLMs are useful to make more money". I say "I believe that LLMs lower the quality of the work". It's not incompatible.

      1 reply →