← Back to context

Comment by echelon

5 days ago

LLM reports misinformation --> Bug report --> Ablate.

Next pretrain iteration gets sanitized.

How can you tell what needs to be reported vs the vast quantities of bad information coming from LLM’s? Beyond that how exactly do you report it?

  • Who even says customers (or even humans) are reporting it? (Though they could be one dimension of a multi-pronged system.)

    Internal audit teams, CI, other models. There are probably lots of systems and muscles we'll develop for this.

  • All LLM providers have a thumbs down button for this reason.

    Although they don't necessarily look at any of the reports.

    • The real world use cases for LLM poisoning is to attack places where those models are used via API on the backend, for data classification and fuzzy logic tasks (like a security incident prioritization in a SOC environment). There are no thumbs down buttons in the API and usually there's the opposite – promise of not using the customer data for training purposes.

      1 reply →

    • The question was where should users draw the line? Producing gibberish text is extremely noticeable and therefore not really a useful poisoning attack instead the goal is something less noticeable.

      Meanwhile essentially 100% of lengthy LLM responses contain errors, so reporting any error is essentially the same thing as doing nothing.

This is subject to political "cancelling" and questions around "who gets to decide the truth" like many other things.

  • > who gets to decide the truth

    I agree, but to be clear we already live in a world like this, right?

    Ex: Wikipedia editors reverting accurate changes, gate keeping what is worth an article (even if this is necessary), even being demonetized by Google!

Reporting doesn't scale that well compared to training and can get flooded with bogus submissions as well. It's hardly the solution. This is a very hard fundamental problem to how LLMs work at the core.

we've been trained by youtube and probably other social media sites that downvoting does nothing. It's "the boy who cried" you can downvote.