← Back to context

Comment by ImPostingOnHN

9 hours ago

The fact that data and instructions are inherently intermixed in most LLMs.

Once either gets into the LLM layer, the LLM can't tell which is which, so one can be treated as the other.

Solutions usually involve offloading some processing to deterministic, non-AI systems which differentiate between the two (like a regular computer program (ignore reflection)), which is the opposite of a "do it all in AI" push from businesses.

The deterministic mixed with LLM approach has been great for me so far. I've been getting a lot of the gains the "do it all with AI" people have been preaching but with far fewer pitfalls. It's sometimes not as fluid as what you sometimes see with the full-LLM-agent setups but that's perfectly acceptable to me and I handle those issues on a case-by-case basis.

  • I'd argue that the moment one cares about accuracy and blast radius, one would natural want to reduce error compounding from a combination of LLM calls (non deterministic) and it's very natural to defer to well tested determinist tools.

    Do one thing and do it well building blocks and the LLM acts a translation layer with reasoning and routing capabilities. Doesn't matter if it's one or an orchestrated swarm of agents.

    https://alexhans.github.io/posts/series/evals/error-compound...

    • Yeah. One of the patterns I've fallen into looks a bit like this:

      1. I have some new task I need/want to do.

      2. For whatever reason, it's not something I want to do myself if I can avoid it.

      3. Have the agent do it the first few times.

      4. After those first few iterations, think about if it's something where the variability in the number of steps needed to complete the task is small enough to just put into a small script or service. If it is, either write the code myself or ask the agent to create draft code based on its own observations of how it did the task those first few times. If it's not, just keep having the agent do it.

      5. A good chunk of the time, most of the task has low variability in what it needs to do except for just one portion. In that case, just use deterministic code for all areas of the program except the high variability area.

      Probably a better word than "variability" for what I'm talking about but I think you get the idea. Spend a lot of tokens upfront so the tokens used later can be minimized when possible.

      EDIT: Formatting.

      1 reply →