Comment by zwaps
16 hours ago
This really doesn’t pan out in practice if you work a lot with these models
And also we know why: effective context depends on inout and task complexity. Our best guess right now is that we are often between 100k to 200k effective context length for frontier, 1m NIHS type models
Agreed, not to mention the m additional cost of chats with more context.