Comment by TZubiri

1 day ago

Maybe it was in the training set.

I think that was Tao's point, that the new proof was not just read out of the training set.

  • I don't think it is dispositive, just that it likely didn't copy the proof we know was in the training set.

    A) It is still possible a proof from someone else with a similar method was in the training set.

    B) something similar to erdos's proof was in the training set for a different problem and had a similar alternate solution to chatgpt, and was also in the training set, which would be more impressive than A)

    • It is still possible a proof from someone else with a similar method was in the training set.

      A proof that Terence Tao and his colleagues have never heard of? If he says the LLM solved the problem with a novel approach, different from what the existing literature describes, I'm certainly not able to argue with him.

      9 replies →

    • Does it matter if it copied or not? How the hell would one even define if it is a copy or original at this point?

      At this point the only conclusion here is: The original proof was on the training set. The author and Terence did not care enough to find the publication by erdos himself

  • The model has multiple layers of mechanisms to prevent carbon copy output of the training data.