← Back to context

Comment by lossolo

1 day ago

> You assume that the system prompt they put on github is the entire system prompt. It almost certainly is not.

It's not about the system prompt anymore, which can leak and companies are aware of that now. This is handled through instruction tuning/post training, where reasoning tokens are structured to reflect certain model behaviors (as seen here). This way, you can prevent anything from leaking.