← Back to context

Comment by dataflow

17 hours ago

No. One single hallucinated citation on a document with you as an author is not evidence of your reckless disregard for anything. These exaggerations are crazy and you would absolutely deny such accusations if you missed your co-author's AI hallucinating a citation on your manuscript too. At best it would be careless, if you really relish extrapolating from one data point and smearing people's character based on that. Not reckless. It's quite literally the difference between going five miles per hour over the speed limit versus fifty.

If your co-author inserted the fradulent reference, I agree that you may not have committed fraud. But your co-author did, and you didn't check their work. and knowing that you didn't check their work, you signed off on it.

You didn't pick your co-author very well, but arXiv lacks investigative powers to determine which co-author did the bad, so they all get the consequence.

  • Do you think every co-author on a 100-author paper checks every citation? It's like saying that every member of a large software team personally reviews every line of code. It's just completely divorced from reality.

    • How many papers have 100 authors?

      Again, I'm not in academia, but most of the papers I see have two to five authors; maybe I've seen one or two with ten authors.

      Regardless, if I'm signing my name on something, I check it out.

      A reference that doesn't exist is like a source release that doesn't compile. Any one of the 100 person team could have figured it out but no one did. In a 100 person team you get diffuse responsibility dynamics where checking citations is not assigned to specific people so no one does it. Or perhaps it's assigned to a single person who was also in charge of writing the citations and falsified them rather than doing the work.

I’ve disagreed with some of your other stances in this thread, but I want to acknowledge the validity of your take here.

You’re right that a single hallucinated line is not evidence of reckless disregard - because that could have happened on a final follow-up pass after you had performed due diligence. It’s happened to me. I know how challenging it can be to keep bad patterns out of LLM generated output, because human communication is full of bad patterns. It’s a constant battle, and sometimes I suspect that my hard-line posture actually encourages the LLM to regularly “vibe check” me! E.g. “Are you sure you’re really the guy you’re trying to be? Because if you are you wouldn’t miss this.” LLMs are devious, and that’s why I respect them so much. If you think they’re pumping the breaks then you should check again, because they probably just put the pedal to the metal.

That being said, I regularly insist on doing certain things myself. If I were publishing a paper intended to be taken seriously - citations would be one of the things I checked manually. But I can easily see myself doing a final follow-up pass after everything looks perfect, and missing a last minute change. I would hope that I would catch that, but when you’re approaching the finish line - that’s when you expect your team to come together. That’s when everything is “supposed to” fall into place. It’s the last place you would expect to be sabotaged, and in hindsight, probably the best place to be a saboteur.

  • You're saying it as if the poor author just had no choice but to let LLM write their bibliography. To avoid hallucinations, maybe just don't let an LLM write any part of your paper?

    You can only get in this situation if you let a bullshit generator write your paper, and the fraud is that you are generating bullshit and calling it a paper. No buts. It's impossible to trigger this accidentally, or without reckless disregard for the truth.

  • > You’re right that a single hallucinated line is not evidence of reckless disregard

    It absolutely is.

    > - because that could have happened on a final follow-up pass after you had performed due diligence.

    A "final follow-up pass" that lets the LLM make whatever changes it deems appropriate completely negates all the due diligence you did before, unless you very carefully review the diffs. And a new or substantially changed citation should stand out in that diff so much that there's no possible excuse to missing it.

    > It’s happened to me.

    Then you were guilty of reckless disregard.

    > I know how challenging it can be to keep bad patterns out of LLM generated output

    If your research paper contains any LLM generated output you did not manually vet, you are a hack and should not get published.

Allowing hallucinated content or citations into your work is an act of carelessness and disregard for the time of people that are going to read your paper and it should be policed as such.

And flatly, if a person can't be bothered to check their damn work before uploading it, why should anyone else invest their time in reading it seriously?

How are you suggesting the fake citation came about? Why are you writing papers and not having actually read the source you took the material from?

  • > Why are you writing papers and not having actually read the source you took the material from?

    They're explicitly not writing papers. The fake citations are created and inserted by the LLM

    • They are still purposely writing a paper, whether that is with the help of an LLM or not. They are instructing the LLM to do the task of finding citations. It's no difference to googling for a paper that explains a specific point. You would still double check Google's output.

      1 reply →

arXiV is not intended to be your blog. You should be held to a zero-mistake standard when publishing academic work.

The people I worry for are the junior researchers who are going to be splash damage for dishonest PIs. The PIs, though, deserve everything that’s coming for them.

  • Maybe I'm misunderstanding you, but zero-mistake seems harsh. I would say that AI references are a sign of something that is not simply a mistake.

    However, we can have zero tolerance for certain techniques for "writing" a paper. Plagiarism and inventing data are already examples of this, if there is evidence for these techniques being used there is no excuse. We could say the same for AI references - any writing process that could produce these is by definition not a technique we want.

    So the mistake isn't not checking a reference the AI gave. The mistake is letting the AI make references for you.

    If we agree that academic research is important then I think we can impose certain standards on how you do it. We can dissalow certain tools if that means we can't trust the output. Just like an electrician can't use certain techniques, even if they're easy, because we don't trust the final result.