Comment by dist-epoch
14 days ago
In a discussion about LLMs you link to a paper from 2023, when not even GPT-4 was available?
And then you say:
> comprehensively disproven
? I don't think you understand the scientific method
14 days ago
In a discussion about LLMs you link to a paper from 2023, when not even GPT-4 was available?
And then you say:
> comprehensively disproven
? I don't think you understand the scientific method
Fair point on the date - the paper was updated October 2024 with Llama-3 and Qwen2.5 (up to 72B), same findings. The v1 to v3 revision is interesting. They initially found personas helped, then reversed their conclusion after expanding to more models.
"Comprehensively disproven" was too strong - should have said "evidence suggests the effect is largely random." There's also Gupta et al. 2024 (arxiv.org/abs/2408.08631) with similar findings if you want more data points.
A paper’s date does not invalidate its method. Findings stay useful only when you can re-run the same protocol on newer models and report deltas. Treat conclusions as conditional on the frozen tasks, criteria, and measurement, then update with replication, not rhetoric.
...or even how fast technology is evolving in this field.