Comment by kam

9 months ago

I think the idea is that if an LLM trained prior to the patent date can reproduce the invention, then either the idea is obvious or there was prior art in the training set; either way the patent is invalid.

> ...if an LLM trained prior to the patent date can reproduce the invention...

Would we even be able to tell if the machine reproduced the invention covered by the claims in the patent?

I (regrettably) have my name on some US software patents. I've read the patents, have intimate knowledge of the software they claim to cover, and see nearly zero relation between the patent and the covered software. If I set a skilled programmer to the task of reproducing the software components that are supposed to be covered by the patents, I guarantee that they'd fail, and fail hard.

Back before I knew about the whole "treble damage thing" (and just how terrible many-to-most software patents are) I read many software patents. I found them to offer no hints to the programmer seeking to reproduce the covered software component or system.

I had similar thoughts before. It's worth thinking about what attorneys will do in response to rejections based on LLMs reproducing ideas. I'm a former patent examiner, and attorneys frequently argue that the examiners showed "hindsight bias" when rejecting claims. The LLM needs to produce the idea without being led too much towards it.

Something like clean-room reverse engineering could be applied. First ask a LLM to describe the problem in a way that avoids disclosing the solution, then ask an independent LLM how that problem could be solved. If LLMs can reliably produce the idea in response to the problem description, that is, after running a LLM 100 times over half show the idea (the fraction here is made up for illustration), the idea's obvious.