Comment by wrsh07
6 days ago
> Each model was run with the recommended hyperparameters and a maximum token limit of 64,000. No models needs more than this number of tokens
I'm a little confused by this. My assumptions (possibly incorrect!): 64k tokens per prompt, they are claiming the model wouldn't need more tokens even for reasoning
Is that right? Would be helpful to see how many tokens the models actually used.
they didn't even do a (non-ml) agentic descent? like have a quicky api that requeries itself generating new context?
"ok here is my strategy here are the five steps", then requery with a strategy or proof of step 1, 2, 3...
in a dfs