← Back to context

Comment by Someone1234

3 days ago

"AI-edited comments" is a very interesting one. Where is the line between a spelling/grammar/tone checker like Grammarly, that at minimum use N-Grams behind the scenes, and something that is "AI" edited? What I am asking is, is "AI" in this context fully featured LLMs, or anything that improves communication via an automated system. I think many people have used these "advanced" spellcheckers for years before Chatgpt et al came on the scene.

I think "generated comments" is a pretty hard line in the sand, but "AI-edited" is anything but clear-cut.

PS - I think the idea behind these policies is positive and needed. I'm simply clarifying where it begins and ends.

You're touching on an important point. More here: https://news.ycombinator.com/item?id=47338091

How do we close the aperture for the lame stuff while opening wider for the good stuff? That is far from clear.

  • Do the guidelines also disallow comments along the lines of "according to <AI>, <blah>"? (I ask this given that "according to a Google search, <blah>" is allowed, AFAIK.)

    • I would lean towards disallowing those. With "According to a Google search ...", someone can ask for specific links (and indeed, people often say to link to those sources to begin with instead of invoking Google). With "According to AI ... " - why would most readers care what the AI thinks? It's not a reliable source! You might as well say "According to a stranger I just met and don't know ..."

      If you're going to say that the AI said X, Y, Z, provide a rationale on why it is relevant. If you merely found X, Y and Z compelling, feel free to talk about it without mentioning AI.

      4 replies →

    • AI is not a source. A Google search result page is not a source. Hopefully, these things help you find a source. If you're posting something you feel the need to source, post the source along with your comment! For example, don't say "according to a Google search, x"... say something like "according to Microsoft's documentation, x" and provide a link to Microsoft Learn page...

    • I don't have a problem with that. First off it's not very common. Second off it can add to a conversation, just as it can with in-person discussions. If you feel like it doesn't, don't upvote and don't reply. There's no value in pretending we're Woodward and Bernstein every time we leave a comment.

    • I think those should be allowed iff the nature of being AI-generated is relevant to the topic of discussion — e.g. if we're talking about whether some model or other can accurately respond to some prompt and people feel inclined to try it themselves.

      1 reply →

    • Citations can be helpful. But AI summaries and Google searches are poor citations because they are not primary sources.

    • We don't want people copy-pasting in comments generally. Summary comments, onlyquote comments (i.e. consisting of a quote and nothing else), duplicate comments are other examples of this. It's not specific to LLMs.

      However, that's probably not critical enough to formally add to the explicit guidelines, so it's probably fine to leave it in the "case law" realm—especially because downvoters tend to go after such comments.

      1 reply →

  • I wasn't sure whether it was an omission or an unintended gap, as the guideline specifically points to "comments". So it seems AI generated/edited posts are fine. Strange, because both can be flagged/downvoted if it was to be left with that.

    • I'm not saying they're all fine, I'm saying we don't yet have any idea of where to make a cut.

      The comments thing is a lot more intimate in the sense that anyone posting comments is inside the house.

  • Please rethink the “edited” bit on accessibility grounds.

    I have a kid with severe written language issues, and the utilisation of speech to text with a LLM-powered edit has unlocked a whole world that was previously inaccessible.

    I would hate to see a culture that discourages AI assistance.

    • That's totally legit and your kid, should they ever take an interest in Hacker News, is welcome here.

      These rules are always fuzzy and there's always a long tail of exceptions. All the more so under turbulent conditions like right now. I wrote more about this elsewhere in the thread, in case it's useful: https://news.ycombinator.com/item?id=47342616.

    • Are you up for sharing details?

      > I would hate to see a culture that discourages AI assistance.

      Mostly I think the push back is about ai assistance in its current form. It can get in the way of communicating rather than assisting. The cost though is mostly borne by the readers and those not using the AI for assistance. I have seen this happen when the ai adds info and thoughts that were tangental to the original author and I think, but I can not verify times where an author seems to try to dig down on the details but seemingly can not.

    • Oh wow. I did not anticipate that, which is embarrassing given that I wrote this just recently:

      https://news.ycombinator.com/item?id=47326351

      Yes, please at least have a carveout for accessibility. I definitely have dictated HN comments in the past, and my flow uses LLMs to clean it up. It works, and is awesome when you're in pain.

    • Since it's mostly a good-faith rule to begin with, it seems easy to add something like, "unless you are using it as an assistive technology for accessibility reasons".

      1 reply →

    • Hear hear. And like many other aspects of accessibility, it will help a huge number of people who may not have any severe issues. e.g. non-native English speakers using LLM-powered edits.

You should use your own words. It might seem that a tool like Grammarly is just an advanced spellcheck, but what it's really doing is replacing your personal style of writing with its own.

It's better to communicate as an individual, warts and all, than to replace your expression with a sanitized one just because it seems "better." Language is an incredibly nuanced thing, it's best for people's own thoughts to come through exactly as they have written them.

  • My elementary school kid came home yesterday and showed me a piece of writing that he was really proud of. It seemed more sophisticated than his typical writing (like, for example, it used the word "sophisticated"). He can be precocious and reads a ton, though, so it was still plausible that he wrote it. I asked him some questions about the writing process to try to tease out what happened, and he said (seemingly credibly) that he hadn't copied it from anywhere or referenced anything. He also said he didn't use any AI tools. After further discussion, I found out that Google Docs Smart Compose (suggested-next-few-words feature) is enabled by default on his school-issued Chromebook, and he had been using it. The structure of the writing was all his, but he said he sometimes used the Smart Compose suggestions (and sometimes didn't). He liked a lot of the suggestions and pressed tab to accept them, which probably bumped up the word choice by several grade levels in some places.

    So yeah, it can change the character of your writing, even if it's just relatively subtle nudges here or there.

    edit: we suggested that he disable that feature to help him learn to write independently, and he happily agreed.

    • To rationalize my gut-feelings on this, I think it comes down to the spectrum between:

      1. A system that suggests words, the child learns the word, determines whether it matches their intent, and proceeds if they like the result.

      2. A system that suggests words, and the child almost-blindly accepts them to get the task over with ASAP.

      The end-results may look the same for any single short document, but in the long run... Well, I fear #2 is going to be way more common.

      12 replies →

    • Oh how I despise these suggestions. You sometimes look for a way to express something and you are on the verge of giving the world something truly original, but as soon as your brain sees the suggestion it goes "oh yeah that fits"

      10 replies →

  • As a non native English speaker my own words wouldnt be in English. If I express myself in English I soon struggle for the right words. On the other hand I think when I read some English text I'm quite capable of sensing the nuances. So it feels when I auto translate my text to English an than read against it again and make some corrections, I can express my thoughts much better.

  • My broken english now officially bumps my comments up instead of down. Sweet.

    • For what it's worth, I had a quick look through your comment history and your English seems just fine to me as a native speaker (at least for informal communication).

      1 reply →

  • >It's better to communicate as an individual, warts and all, than to replace your expression with a sanitized one just because it seems "better."

    It is definitely not true that it is better for a poster to communicate like an individual when it comes to spelling and grammar. People ignore posts that have poor grammar or spelling mistakes, and communications that have poor grammar are seen as unprofessional. Even I do it at a semi-subconscious level. The more difficult or the more amount of attention someone has to pay to understand your post, the less people will be willing to put in that effort to do so.

    • Exactly. Tell that to whoever is grading your next paper, or reviewing your resume, or watching your presentation. People are judged by their linguistic ability even in cases where it shouldn't matter. It's a well known heuristic bias. It's no surprise that many of the people here denying it are themselves quite literate.

  • Books and newspapers have had editors for centuries. It is just code review for the written word.

    [It looks like MS Word 97 had the ability to detect passive voice as well, so we're talking 30 year old technology there that predates LLMs -- how far down the Butlerian Jihad are we going with this?]

    • Editors are mostly tasked with maintaining a consistent style and standard.

      There is no need for that here beyond maybe spellcheck. Use your own thoughts, voice, and words.

      10 replies →

  • I was just re-reading the passage from Plato's "The Phaedrus" on writing & the "art" of the letter for an essay I'm working on, and your remark is salient for this discussion on LLM-style AI and social media at large.

  • Precisely. As I wrote in my assessment of AI for my workplace;

    "Your unique human voice is more valuable than a thousand prompt-driven LLM doggerels."

  • That's true, but on the flip side I regularly get downvoted because my English is not the best, so say it mildly. So, now I need to be really careful, to a) write in a good English or b) not to be recognised as an LLM corrected version of my English. Where is the line? I shouldn't be downvoted for my English I think, but that is the reality.

    Edit: I already got downvoted. :-) Sure, no one can tell exactly why. Maybe the combination of bad English _and_ talking sh*ce isn't ideal at all. :-D Anyways, I have enough karma, so I can last quite a while..

    • It goes both ways.

      The quality of my writing varies (based on my mood as much as anything else, I suppose), but when it is particularly good and error-free then I often get accused of being a bot.

      Which is absurd, since I don't use the bot for writing at all.

    • > I shouldn't be downvoted for my English I think, but that is the reality.

      How do you know? Is it possible the downvoters just didn't like what you said?

      2 replies →

  • > It's better to communicate as an individual, warts and all, than to replace your expression with a sanitized one just because it seems "better." Language is an incredibly nuanced thing, it's best for people's own thoughts to come through exactly as they have written them.

    This is the opposite of how language works. You want people to understand the idea you're trying to communicate, not fixate on the semantics of how you communicated. Language is like fashion - you only want to break the rules deliberately. If AI or an editor or whatever changes your writing to be more clear and correct, and you don't look at it and say "no, I chose that phrasing for a reason" then the editor's version is much more likely to be understood correctly by the recipient.

  • I'm not sure I agree with this. I don't really want to see someone else's stylistic "warts".

    I just want clean, easy-to-read content and I don't care about the person who wrote it. A tool like Grammarly is the difference between readable and unreadable (or understandable and understandable) for many people.

    • You could run the comments everyone else posts through an AI tool and ask it to rephrase it so that it is clean, and easy-to-read.

      You could even write a plugin for your favorite web browser to do that to every site you visit.

      It seems hard to achieve the inverse that is (would you rather I use i.e.?) rewrite this paragraph as the original author did before they had an AI re--write it to make it clean, (--do you like oxford commas, and em/en dashes! Just prompt your AI) and easier to read

      3 replies →

  • I disagree. HN is going to bury my raw unedited tirade of a comment about those fucking morons that couldn't code their way out of a paper bag. If I send a comment to ChatGPT and open up the prompt with "this poster is a fucking dumbass, how do I tell them this" and use that to get to a well reasoned response because that's the tool we have available today, we're all better off.

    The guidelines state:

    > Be kind. Don't be snarky. Converse > Edit out swipes. > Don't be curmudgeonly.

    On the best of days I manage to follow the rules, but I'm only human. If I run my comment through ChatGPT to try and help me edit out swipes on the bad days, that's not ok?

    I'm not using ChatGPT to generate comments, but I've got the -4 comments to show that my "thoughts exactly as they have written them" isn't a winning move.

    • If you see an incompetent coder and wish to communicate that the person responsible is a "fucking moron/dumbass", the tone with which you do so is not the problem. Tell us what is wrong with the code, as objectively as possible. That's what the guidelines are trying to convey.

    • The guidelines don't say anything about not posting something because an LLM told you that you shouldn't...

  • But the problem is that people with poor written language / english skills are 'competing' with people who have superb skills in this domain.

    There are people here who sit at a desk all day banging out multipage emails for work who decide to write posts of a similar linguistic calibre for funsies.

    Meanwhile you have someone in a developing country who just got off a brutal twelve hour shift doing manual labour in the sun who wants to participate in the conversation with an insightful message that they bang-out on a shitty little cellphone onscreen keyboard while riding on bumpy public transit.

    You could have a great idea and express it poorly and be penalized for doing so here while someone could have a blah idea expressed excellently and it's showered in replies despite being in some metrics (the ones I think are most important) worse than the other post.

    What's the solution for that?

    • > You could have a great idea and express it poorly and be penalized for doing so here while someone could have a blah idea expressed excellently and it's showered in replies despite being in some metrics (the ones I think are most important) worse than the other post.

      I absolutely do not understand this comment. Are you saying that posting is competitive and that comments have "metrics"?

      1 reply →

I think that the line between A"I" editing to fix grammar or to translate from a different native language and A"I" editing by using an LLM is one of those things that's very hard to unambiguously encode in written guidelines, but easy to intuitively understand using common sense, in the vein of I know it when I see it.

https://en.wikipedia.org/wiki/I_know_it_when_I_see_it

> Where is the line between a spelling/grammar/tone checker like Grammarly

For me, the line is precisely at the point where a human has something they want to say. IMO - use the tools you need to say the thing you want to say; it's fine. The thing I, and many others here, object to is being asked to read reams of text that no-one could be bothered to write.

On a technical level, you can really only guard against changing your semantics and voice - if you're letting software alter the meaning, or meanings, you intend, and use words you don't normally use, it's probably too far.

This is probably ok:

>> On a technical level, you can really only guard against software that changes your semantics or voice. If you're letting it alter the meaning (or meanings) you intend, or if it starts using words you would never normally use, then it's gone too far.

This is probably too far:

>>> On a technical level, it's important to recogn1ize that the only robust guardrail we can realistically implement is one that prevents modifications to core semantics or authorial voice. If you're comfortable allowing the system to refine or rephrase the precise meanings you originally intended — or if it begins incorporating vocabulary that doesn't align with your typical linguistic patterns — then you've likely crossed a meaningful threshold where the output no longer fully represents your authentic intent.

Something to consider is that you can analyze your own stylometric patterns over a large collection of your writing, and distill that into a system of rules and patterns to follow which AI can readily handle. It is technically possible, albeit tedious, to clone your style such that it's indistinguishable from your actual human writing, and can even icnlude spelling mistakes you've made before at a rate matching your actual writing.

AI editing is weird, though. Not seeing a need, unless English isn't your native language.

I think there's a pretty clear gap between editing for grammar/spelling and editing for tone.

  • How so and why? I know plenty of people whose writing naturally carries a tone that they don't intend. I often help them to change their wording to be less confrontational or seemingly sarcastic when it isn't meant to be. Would you say it is wrong for them to get assistance to get the tone they intend rather than the one they would tend to write?

    • It's the difference between correctness and tone/character/semantics (tone and character do affect semantics). We need to do things we don't quite mean in subjective spaces, to learn. Developing yourself is wonderful, but presenting a writing style that does not yet represent your learned tone feels disingenuous to the reader and harms the tone of the whole conversation. Using LLMs to iterate might help you learn, but use that tool privately, or with friends/family/mentors. With others, simply make your mistakes.

      To be clear, I also think you shouldn't rely on auto-correction or LLMs for correctness (they are great for identifying your mistakes, but I think you should then fix the mistakes yourself, to develop your brain). It's just that "assisted" correctness isn't misleading/harmful in the way that "assisted" tone/character/semantics are.

Trying to lawyer this is the wrong approach. When in doubt: don't.

  • That feels very uncharitable.

    When a policy is introduced to seemingly guard against new problems, but happens to be inadvertently targeting preexisting and common technology, I don't feel like it is "lawyering" it to want clarity on that line.

    For example, it could be argued this forbids all spellcheckers. I don't think that is the implied intent, but the spectrum is huge in the spellchecker space. From simple substitutions + rule-based grammar engines through to n-grams, edit-distance algorithms, statistical machine translation, and transformer-based NLP models.

I think the only practical litmus test here is whether you can stand by the text as your own words. It’s not like we have someone looking over commenters’ shoulders as they type.

Ultimately, this comes down to people making a good-faith judgment about how much AI was involved, whether it was just minor grammatical fixes or something more substantial. The reality is that there isn’t really a shared consensus on exactly where that line should be drawn.

Grammarly use is outright prohibited by this; AI-edited writing is no longer writing that you hold personal and exclusive responsibility for having written. Consider Stephen Hawking’s voice box generator. While the sounds produced were machine-assisted, the writing was his alone. If you find yourself unable to participate in this web forum without paying a proofreader (in time, money, or cycles) to copy-edit your writing, then you’re not welcome on HN as a participant.

  • > If you find yourself unable to participate in this web forum without paying a proofreader (in time, money, or cycles) to copy-edit your writing, then you’re not welcome on HN as a participant.

    You forgot the /s ?

    • It’s not sarcasm. If you feel if I have misunderstood the intent of the guideline we’re discussing — “Don’t post generated/AI-edited comments”, as the title currently reads, then I’m happy to discuss further. (I often make logical negation errors that I miss in proofing, so it’s possible I slipped up, too!)

      2 replies →

Finding it more refreshing these days when reading text with broken grammar, incorrect use of pronouns, etc. especially for HN, the human connection is more palpable. It’s rarely so bad that it’s not understandable

I saw a similar conversation somewhere about some project saying they don't allow AI generated code.

It was asked that if "AI Generated Code" is just code suggested to you by a computer program, where does using the code that your IDE suggests in a dropdown? That's been around for decades. Is it LLM or "Gen AI" specific? If so, what specific aspect of that makes one use case good and one use case bad and what exactly separates them?

It's one of those situations where it seems easy to point at examples and say "this one's good and this one's bad", but when you need to write policy you start drowning in minutia.

  • Projects cannot allow AI generated code if they require everything to have a clear author, with a copyright notice and license.

    IDE code suggestions come from the database of information built about your code base, like what classes have what methods. Each such suggestion is a derived work of the thing being worked on.

    • That is not correct because it hasn't been tested in court. In past decisions about who owns the output generated by a computer program the owner has been the operator of the program. You own your Word documents and Photoshopped images. There is good reason to believe that LLM output where you provided the prompt would also fit under that umbrella. We are still waiting for that to be tested in court.

      1 reply →

  • Nobody is actually confused about what AI generated code means in those cases, they're just trying to be argumentative because they don't like the rules

Your comment is one of semantics. Worth discussing if we're talking a truly hard line rule rather than the spirit of the rule.

I benefit from my phone flagging spelling errors/typos for me. Maybe it uses AI or maybe it uses a simple dictionary for me. Maybe it might even catch a string of words when the conjunction isn't correct. That's all fair game, IMO. But it shouldn't be rewriting the sentence for me. And it shouldn't be automatically cleaning up my typos for me after I've hit "reply". That's on me.

I don’t think it’s really necessary to play Captain Nitpick over spell-check or whatever. You know what is meant.

I caught myself structuring a comment like an LLM on another site. It's expected that people who chat heavily to LLMs will start to mirror their styles.

I agree on the editing. We use these things all the time - chances are many of you are using it right now as you type on your phone and it checks your spelling for you.

By the same token, what if I have a human editor help me out? What if we go back and forth on how to write something, including spelling, grammar, tone, etc. For example, my wife occasionally asks me to review her messages before sending them because she thinks I speak well and wants to be understood correctly.

The problem is that we are punishing the technology, not the result. Whether it's a human or an LLM that acts as your editor should be irrelevant; what matters is that you are posting your own work and not someone else's. My wife having me write all of her messages for her would be just as dishonest as her having an LLM write all of her messages for her if she always presented them as her own writing. But if she writes the copy and I provide suggests for changes, what's the harm in that? And why should it matter if it's a human or an LLM that provides that assistance?

i don't care if someone has bad grammar, i want to hear their thoughts as they came up with them, we're all intelligent beings and can parse the meaning behind what you write.

i type my comments without capitalization like i'm typing into some terminal because i'm lazy and people might hate it but i'm sure they prefer this to if i asked an LLM to rewrite what i type

your writing style is your personality, don't let a robot take it away from you

  • I, on the other hand, find incorrect grammar mildly annoying, especially when it's due to laziness. It distracts from the thoughts being conveyed. I appreciate when people take the time to format comments as correctly as they're able.

    In fact, I'd argue that lazy commenting is the real problem, which has now been supercharged by LLMs.

ML based word or phrase editing is hardly a problem any more than pre-AI spellcheckers were. AI sentence and paragraph manufacturing is a problem and everyone knows the difference between that slop and a spellchecker. No one cares if your editor does inline spellchecking or even word autocomplete. What they care about is slop and word at a time spelling or phrase grammar checking are harmless.