Comment by guelo 5 months ago How did you chose the 8192 token thinking budget? I've often seen Deepseek R1 use way more than that. 1 comment guelo Reply freediver 5 months ago Arbitrary, and even with this budget it is already more verbose (and slower) overall than all the other thinking models - check tokens and latency in the table.
freediver 5 months ago Arbitrary, and even with this budget it is already more verbose (and slower) overall than all the other thinking models - check tokens and latency in the table.
Arbitrary, and even with this budget it is already more verbose (and slower) overall than all the other thinking models - check tokens and latency in the table.