Comment by BoorishBears
2 years ago
The difference is UX: Are you going to have your user work around poor prompting by giving examples with every request?
Instead of a UI that's "Describe what you want" you're going to have "Describe what you want and give me some examples because I can't guarantee reliable output otherwise"?
Part of LLMs becoming more than toy apps is the former winning out over the latter. Using techniques like chain of thought with carefully formed completions lets you avoid the awkward "my user is an unwilling prompt engineer" scenarios that pop up otherwise.
> Are you going to have your user
What fucking user, man? Is it not painfully clear I never spoke in the context of deploying applications?
Your issues with this level of prefilling in the context of deployed apps ARE valid but I have no interest in discussing that specific use case and you really should have realized your arguments were context dependent and not actual rebuttals to what I claimed at the start several comments ago.
Are we done?
I thought we were done when I demonstrated GPT 4 can continue a completion contrary to your belief, but here you are throwing a tantrum several comments later.
> GPT 4 can continue a completion contrary to your belief
When did I say that? I said they work differently. Claude has nothing in between the prefill and the result, OpenAI has tokens between the last assistant message and the result, this makes it different. You cannot prefill in OpenAI, Claude's prefill is powerful as it effectively allows you to use it as general completion model, not a chat model. OpenAI does not let you do this with GPT.
4 replies →