Comment by dvt
1 day ago
> Is anyone doing it this way?
I'm working on a way of invoking tools mid-tokenizer-stream, which is kind of cool. So for example, the LLM says something like (simplified example) "(lots of thinking)... 1+2=" and then there's a parser (maybe regex, maybe LR, maybe LL(1), etc.) that sees that this is a "math-y thing" and automagically goes to the CALC tool which calculates "3", sticks it in the stream, so the current head is "(lots of thinking)... 1+2=3 " and then the LLM can continue with its thought process.
Cold winds are blowing when people look at LLMs and think "maybe an expert system on top of that?".
I don't think it's "on top"? I think it's an expert system where (at least) one of the experts is an LLM, but it doesn't have to be LLMs from bottom to top.
On the side, under, wherever. The point is, this is just re-inventing past failed attempts at AI.
7 replies →
Definitely an interesting thought to do this at the tokenizer level!