Comment by imiric
15 hours ago
Ah, that's good to know, thanks.
But then why is there compounding token usage in the article's trivial solution? Is it just a matter of using the cache correctly?
15 hours ago
Ah, that's good to know, thanks.
But then why is there compounding token usage in the article's trivial solution? Is it just a matter of using the cache correctly?
Cached tokens are cheaper (90% discount ish) but not free
Also, unlike OpenAI, Anthropic's prompt caching is explicit (you set up to 4 cache "breakpoints"), meaning if you don't implement caching then you don't benefit from it.
thats a very generous way of putting it. Anthropic's prompt caching is actively hostile and very difficult to implement properly.