And Outlier/Scale, which was bought by Meta (via Scale), had many IMO-required Math AI trainer jobs on LinkedIn. I can't find those historical ones though.
I'm just one piece in the cog and this is an anecdote, but there was a huge upswing in IMO or similar RLHF job postings over the past 6mo-year.
https://work.mercor.com/jobs/list_AAABljpKHPMmFMXrg2VM0qz4
https://benture.io/job/international-math-olympiad-participa...
https://job-boards.greenhouse.io/xai/jobs/4538773007
And Outlier/Scale, which was bought by Meta (via Scale), had many IMO-required Math AI trainer jobs on LinkedIn. I can't find those historical ones though.
I'm just one piece in the cog and this is an anecdote, but there was a huge upswing in IMO or similar RLHF job postings over the past 6mo-year.
I would fully expect every IMO participant grinds IMO problems for months before the competition.
I don't know why people hold training a model on like material as a negation of it's ability.
It shows models need RL for any new domain/level of expertise, which is contrary to what the marketers claim about LLMs and potential for AGI.