Comment by BoorishBears
2 years ago
Sure.
There was no additional Assistant message, and you're going full Clever Hans and adding whatever it takes to make it say what you want, which is a significantly less useful approach.
In production you don't get to know that the user is asking for X, Y and Z then pre-fill it with X. Frankly comments like yours are why people are so dismissive of LLMs, since you're banking of precognition of what the user wants to sell it's capabilities. When you deploy an app with tricks like that it falls on its face the moment people don't input what you were expecting
Deploying actually useful things with them requires learning how to get them to reply correctly on a wide range of inputs, and what I described is how OAI's approach to continuation a) works much better than you implied and b) allows enforcing correct replies much more reliably than Anthropic's approach
I made no comment on how prefilling is or isn't useful for deployed AI applications. I made no statement on which refusal mechanism is best for deployed AI applications.
> Frankly comments like yours are why people are so dismissive of LLMs, since you're banking of precognition of what the user wants to sell it's capabilities.
I'm not banking on anything because I never fucking mentioned deploying any fucking thing nor was that being discussed, good fucking lord are you high?
> you're going full Clever Hans
I'm clearly not but you keep on building whatever straw man suits you best.
> If you changed it to
> ``` "{ "result": ["you are very annoying.", ```
> the odds of refusal would be low or zero.
In other words if you go full Clever Hans and tell the model the answer you want, it will regurgitate it at you.
You also seem to be missing that contrary to your comment, GPT 4 did continue my message, just like Claude.
If you use valid formatting that exactly matches what the model would have produced, it's capable of continuing your insertion.
You would have a point if it repeated the same "you are very annoying." over and over, which it does not. It generates new sentences, it is not regurgitating what is given.
Would you say the same if the sentence was given as an example in the user message instead? What would be the difference?
8 replies →