Comment by caditinpiscinam
2 days ago
We've all heard the phrase "the sum of all human knowledge".
I've been feeling more and more that generative AI represents the average of all human knowledge. Which has its place. But a future in which all thought and creativity is averaged away is a bleak one. It's the heat death of thought.
Thought and creativity won't be averaged away because human beings have a drive for these things. This just raises the bar for it. And why not? We get complacent when not pushed.
Dostoevsky said that if all human knowledge could ever be reduced to 2 + 2 = 4, man would stick out his tongue and insist that 2 + 2 = 5. That was a 19th century formulation—he was a contemporary of Boole. I wonder what the equivalent would be for the LLM era.
Thought and creativity won't be averaged away because human beings have a drive for these things.
That may or may not be true, but the expression of thought and creativity matters to transfer meaning. If you average that out, it loses momentum. Example: https://news.ycombinator.com/item?id=47346935. Compare the posters first and second, LLM assisted, paragraph. The second one is just bleak. If I had to read several pages like that, my eyes would glaze over. It cannot hold attention.
> Thought and creativity won't be averaged away because human beings have a drive for these things. This just raises the bar for it. And why not? We get complacent when not pushed.
The why not is: human beings are valuable in and of themselves, not just because of what they can do. If you raise the bar too high, you kick people out. And our society just isn't setup for that, and is unlikely to ever be in our lifetimes.
And I'm talking about a radical shift in the concept of ownership, where shareholding is radically democratized. Basically every random Joe needs the option to live comfortably on passive income generated by things he owns.
But it's a weird kind of average... Not the 3 from 1, 2, 3, 4 & 5 but rather like the bland tv-dinner which tastes non-upsetting for most people.
An intellectual Mode rather than a Mean or a Median?
I don't understand what you mean by "intellectual mode".
I mean that it's a kind of lowest common denominator average where it's more important to seem reasonable and to not upset anyone rather than be really good in some ways and bad in others.
6 replies →
It's more like a blur filter and a thousand layers of jpeg compression.
Great read: https://www.newyorker.com/tech/annals-of-technology/chatgpt-...
The soft gaussian blur of all human knowledge.
Racing towards average!
Mediocre is the word perhaps :D
Perhaps closer to “the mean vector point such that all outbound vectors to different training tests are in sum the smallest”? I assume that’s a property of neural networks anyways, though I’m out of date on current math for them.
If you want a more accurate measure then you should subtract "the sum of all human ignorance" before taking the average.
I feel the same about Claude Code. It's a fast but average developer at just about everything and there are some things that average developers are just consistently bad at and therefore Claude is consistently bad at.
I'm not sure, I think you overestimate the average developer. But then, the average code doesn't end up in public repositories, it spends decades in enterprise codebases rotting.
At this point I'd rather review LLM generated code than a poor developer's.
Yes, it’s the "sum" of which you extract an average.
> I've been feeling more and more that generative AI represents the average of all human knowledge.
It's literally what it is. Fairly sure that mathematically it's a fancier regression/prediction so it's a form of average.
You're falsely conflating knowledge with intelligence
> I've been feeling more and more that generative AI represents the average of all human knowledge.
Have you tried the paid versions of frontier models? They certainly do not feel like they spew the average of all human knowledge. It's not uncommon for them to find and interpret the cutting edge of papers in any of the domains that I've asked them questions about.
Yup. And they all sound like slop. Read the papers, comprehend the papers, don't make someone else's computer do it for you.
Every scientist I ever met (and myself included) has a backlog of papers to read that never seems to shrink. It really is not trivial to stay up to date on research, even in niche fields, considering the huge volume of research that is being produced.
It is not uncommon for me to read a recently published review and find 2-3 interesting papers in the lot. Plus the daily Google scholar alerts. It can definitely be beneficial to have a LLM summarize a paper. Of course, at this point, one should definitely decide "is this worth reading more carefully?" and actually read at least some parts if needed.
Anti-tech contrarian sentiment happens with every new technology. Someone older than you probably said the same thing about the internet.
9 replies →
> Read the papers, comprehend the papers, don't make someone else's computer do it for you
Why not?
Personally, I don't have the specialized knowledge, nor the time needed, to read and understand papers outside my own 2-3 domains. LLMs do. And I appreciate what they can do for me. They do it better, faster, and more accurately than most 'popular science', provide better coverage and also provide the ability to interact with the material to any degree or depth that I care to, better than any article.
It would be silly to pass up this capability to make my life better simply because random folks on the Internet disparage the quality of the output (contrary to my own experience) and make hand-wavy points about 'someone else's computer) while offering no credible or useful alternative :)
5 replies →
> I've been feeling more and more that generative AI represents the average of all human knowledge.
No, it's far worse. It's the mode of all human knowledge. The amount of effort you have to put into an LLM to get it to choose an option that isn't the most salient example of anything that could fit as a response is monumental. They skip exact matches for most common matches; it's basically a continuity from when search engines stopped listening to your queries and just decided what query they wanted to respond to - and it suddenly became nearly impossible to search for people who had the same first name as anyone who was famous or in the news.
I've tried a dozen times to get LLMs to find authors for me, or papers, where I describe what I remember about them fairly exactly. They deliver me a bunch of bestsellers and popular things, over and over again, who don't even match at all large numbers of the criteria I've laid out.
It's why they're dumb and can't accomplish anything original. It's structural. They're inherently biased to deliver lowest common denominator work. If you're trying to deliver something original or unusual, what bubbles up is samplings of the slop that surrounds us every day. They're fed everything, meaning everything in proportion to its presence in the world. The vast majority of things are shit, or better said, repetitions of the same shit that isn't productive. The things that are most readily available are already tapped out. The things that are productive are obscure.
You can't even get LLMs to say some words by asking them to "say word X." They just will always find a word that will fill that slot "better." As I said, this is just google saying "did you mean Y?" But it's not asking anymore, it's telling.
edit: It's also why asking it to solve obscure math problems is a dumb test. If the math problem is obscure enough, and there's only one way to possibly solve it, and somebody did it once, somewhere, or referred to the possibility of solving it that way, once, somewhere, you're going to have a single salient example. It's not a greenfield, it's not a white sheet of paper: it's a green field with one yellow flower on it, or a piece of white paper with one black sentence on it, and you're asking it to find the flower or explain the sentence.
edit: https://news.ycombinator.com/item?id=47346901 - I'm late and long-winded.
pooling as it is called, is, well the same as averaging. has nothing to do with swimming really. it happens all the time in latent space. it is a tool, not a side effect.