Slacker News Slacker News logo featuring a lazy sloth with a folded newspaper hat
  • top
  • new
  • show
  • ask
  • jobs
Library
← Back to context

Comment by Der_Einzige

2 years ago

I hope that the long context length models start getting better. Claude 1 and GPT-4-128K both struggle hard once you get past about 32K tokens.

Most of the needle in a haystack papers are too simple of a task. They need harder tasks to test these long context length models for if they are truly remembering things or not.

0 comments

Der_Einzige

Reply

No comments yet

Contribute on Hacker News ↗

Slacker News

Product

  • API Reference
  • Hacker News RSS
  • Source on GitHub

Community

  • Support Ukraine
  • Equal Justice Initiative
  • GiveWell Charities