Comment by pessimizer

2 days ago

> I've been feeling more and more that generative AI represents the average of all human knowledge.

No, it's far worse. It's the mode of all human knowledge. The amount of effort you have to put into an LLM to get it to choose an option that isn't the most salient example of anything that could fit as a response is monumental. They skip exact matches for most common matches; it's basically a continuity from when search engines stopped listening to your queries and just decided what query they wanted to respond to - and it suddenly became nearly impossible to search for people who had the same first name as anyone who was famous or in the news.

I've tried a dozen times to get LLMs to find authors for me, or papers, where I describe what I remember about them fairly exactly. They deliver me a bunch of bestsellers and popular things, over and over again, who don't even match at all large numbers of the criteria I've laid out.

It's why they're dumb and can't accomplish anything original. It's structural. They're inherently biased to deliver lowest common denominator work. If you're trying to deliver something original or unusual, what bubbles up is samplings of the slop that surrounds us every day. They're fed everything, meaning everything in proportion to its presence in the world. The vast majority of things are shit, or better said, repetitions of the same shit that isn't productive. The things that are most readily available are already tapped out. The things that are productive are obscure.

You can't even get LLMs to say some words by asking them to "say word X." They just will always find a word that will fill that slot "better." As I said, this is just google saying "did you mean Y?" But it's not asking anymore, it's telling.

edit: It's also why asking it to solve obscure math problems is a dumb test. If the math problem is obscure enough, and there's only one way to possibly solve it, and somebody did it once, somewhere, or referred to the possibility of solving it that way, once, somewhere, you're going to have a single salient example. It's not a greenfield, it's not a white sheet of paper: it's a green field with one yellow flower on it, or a piece of white paper with one black sentence on it, and you're asking it to find the flower or explain the sentence.

edit: https://news.ycombinator.com/item?id=47346901 - I'm late and long-winded.