← Back to context

Comment by astrange

9 hours ago

Claude's system prompt is still incredibly long and probably hurting its performance.

https://github.com/asgeirtj/system_prompts_leaks/blob/main/A...

They ain't called guard rails for nothing! There's a whole world "off-road" but the big names are afraid of letting their superintelligence off the leash. A real shame we're letting brand safety get in the way of performance and creativity, but I guess the first New York Times article about a pervert or terrorist chat bot would doom any big name partnerships.

  • Anthropic's entire reason for being is publishing safety papers along the lines of "we told it to say something scary and it said it", so of course they care about this.

    • I can't stand this myopic thinking.

      Do you want to learn "oh, LLMs are capable of scheming, resisting shutdown, seizing control, self-exfiltrating" when it actually happens in a real world deployment, with an LLM capable of actually pulling it off?

      If "no", then cherish Anthropic and the work they do.

      7 replies →