Comment by msp26
2 years ago
That 200k context needs some proper testing. GPT-4-Turbo advertises 128k but the quality of output there goes down significantly after ~32k tokens.
2 years ago
That 200k context needs some proper testing. GPT-4-Turbo advertises 128k but the quality of output there goes down significantly after ~32k tokens.
Read the article, it's addressed with charts.
I did but I want more independent testing than just QA performance by position.