Comment by nblgbg

2 years ago

The issue of hallucinations won't be solved with the RAG approach. It requires a fundamentally different architecture. These aren't my words but Yann LeCun's. You could easily understand if you spend some time playing around. The autoregressive nature won't allow the LLMs to create an internally consistent model before answering the question. We have approaches like Chain of Thought and others, but they are merely band-aids and superficially address the issue.

If you build a complex Chain if Thought style Agent and then train/finetune further by reinforcement learning with this architecture then it is not a band-aid anymore, it is an integral part of the model and the weights will optimize to make use of this CoT ability.

  • It's been 3.5 years since GPT-3 was released, and just over a year since ChatGPT was released to the public.

    If it was possible to solve LLM hallucinations with simple Chain-of-Thought style agents, someone would have done that and released a product by now.

    The fact that nobody has released such a product, is pretty strong evidence that you can't fix hallucinations via Chain-of-Thought or Retrieval-Augmented Generation, or any other band-aid approaches.

    • I agree: but I just wanted to say that there are specific subdomains where you can mitigate some of these issues.

      For example, generating json.

      You can explicitly follow a defined grammar to get what will always be a valid json output.

      Similarly, structured output such as code can be passed to other tools such as compilers, type checkers and test suites to ensure that at a minimum the output you selected passes some minimum threshold of “isn’t total rubbish”.

      For unstructured output this a much harder problem, and bluntly, it doesn’t seem like there’s any kind of meaningful solution to it.

      …but the current generation of LLMs are driven by probabilistic sampling functions.

      Over the probability curve you’ll always get some rubbish, but if you sample many times for structure and verifiable output you can, to a reasonable degree, mitigate the impact that hallucinations have.

      Currently that’s computationally expensive, to drive the chance of error down to a useful level, but compute scales.

      We may seem some quite reasonable outputs from similar architectures wrapped in validation frameworks in the future, I guess.

      …for, a very specific subset of types of output.

      2 replies →