Comment by kenjackson

6 months ago

I think this is both right and wrong. There was a good book that came out probably 15 years ago about how technology never stops in aggregate, but individual technologies tend to grow quickly and then stall. Airplane jets were one example in the book. The reason why I partially note this as wrong is that even in the 70s people recognized that supersonic travel had real concrete issues with no solution in sight. I don't think LLMs share that characteristic today.

A better example, also in the book, are skyscrapers. Each year they grew and new ones were taller than the ones last year. The ability to build them and traverse them increased each year with new technologies to support it. There wasn't a general consensus around issues that would stop growth (except at more extremes like air pressure). But the growth did stop. No one even has expectations of taller skyscrapers any more.

LLMs may fail to advance, but not because of any consensus reason that exists today. And it maybe that they serve their purpose to build something on top of them which ends up being far more revolutionary than LLMs. This is more like the path of electricity -- electricity in itself isn't that exciting nowadays, but almost every piece of technology built uses it.

I fundamentally find it odd that people seem so against AI. I get the potential dystopian future, which I also don't want. But the more mundane annoyance seems odd to me.

> even in the 70s people recognized that supersonic travel had real concrete issues with no solution in sight. I don't think LLMs share that characteristic today

I think they pretty strongly do

The solution seems to be "just lower your standards for acceptable margin of error to whatever the LLM is capable of producing" which should be concerning and absolutely unacceptable to anyone calling themselves an Engineer

  • 99% or more of software developers behave in ways that would be inconceivable in actual engineering. That's not to say there aren't software engineers, but most developers aren't engineers and aren't held to that standard.

    • Code is not physical. While computation errors can have real effects, a lot of orgs and people are resilient about them.

  • “Increasing Success by Lowering Expectations” That is from Despair Inc. This was obviously meant to be funny by them, now it looks like the state of play.

  • > absolutely unacceptable to anyone calling themselves an Engineer

    Isn’t that exactly what engineers do? Even very strong bridges aren’t designed to survive every possible eventuality.

    • No

      I'm talking about engineering a bridge for 50 cars that collapses at 51, not engineering a bridge for 500 cars that is only expected to get 50

      Engineering does require tradeoffs of course. But that's not what the minimum possible quality is

    • That's what a "margin of error" is. The margin of error of a bridge is predictable thanks to well-established techniques of physical analysis.

      An LLM system, on the other hand, can fail because you moved some punctuation around.

      1 reply →

> The reason why I partially note this as wrong is that even in the 70s people recognized that supersonic travel had real concrete issues with no solution in sight. I don't think LLMs share that characteristic today.

The fundamental problem has already been mentioned: Nobody can figure out how to SELL it. Because few people are buying it.

It's useful for aggregation and summarization of large amounts of text, but it's not trustworthy. A good summary decreases noise and amplifies signal. LLMs don't do that. Without the capability to validate the output, it's not really generating output of lasting value. It's just a slightly better search engine.

It feels like, fundamentally, the primary invention here is teaching computers that it's okay to be wrong as long as you're convincing. That's very useful for propaganda or less savory aspects of business, but it's less useful for actual communication.

  • > Nobody can figure out how to SELL it. Because few people are buying it.

    Just picking one company who basically just does AI, OpenAI. They reported it has 20 million PAID subscribers to ChatGPT. With revenue projected above $12b dollars (https://www.theverge.com/openai/640894/chatgpt-has-hit-20-mi...).

    I think what you meant to say is that costs are high so they can't generate large profits. but saying that they can't figure out how to sell it seems absurd. Is it Netflix level of subscribers, no. But there can't be more than a couple of hundred products that have that type of subscription reach.

    • Ok but isn’t 20 million subscribers out of what, 800 million or 1 billion monthly users or whatever they’re claiming, an absolutely abysmal conversion rate? Especially given that the industry and media have been proclaiming this as somewhere between the internet and the industrial revolution in terms of impact and advancement? Why can they not get more than 3% of users to convert to paying subscribers for such a supposedly world changing technology, even with a massive subsidy?

      10 replies →

  • In my companies, AI subscriptions and API access are now the biggest costs after salaries and taxes. Don't know what makes you think these services aren't attracting paid customers?

> even in the 70s people recognized that supersonic travel had real concrete issues with no solution in sight. I don't think LLMs share that characteristic today

I hate to dogpile on this statement but I can think of two major issues right now:

* Small context windows, and serious degradation when pushing the limits of existing context windows. A human can add large amounts of state to their "context window" every day.

* Realtime learning. My humans get smarter every day, especially in the context of working with a specific codebase.

Maybe the AI companies will figure this out, but they are not "same technique more processor power" kinds of problems.

There are sound math reasons for skyscrapers topping out, mostly due to elevator capacity and the inability to effectively get people in and out of the floorspace as you go past a few hundred ft. There's no construction engineering reason you can't go taller - the Burj Khalifa, for example, is three times taller than a typical Western major city skyscraper - it just doesn't make economic sense unless you're a newly rich nation looking to prove a point.

  • Economic Concrete construction (what China specializes in) typically tops out at 30-40 floors, so the vast majority of buildings in Asia are that height, a sweet spot so to speak especially for residential (even in limited space HK).

>I think this is both right and wrong. There was a good book that came out probably 15 years ago about how technology never stops in aggregate, but individual technologies tend to grow quickly and then stall. Airplane jets were one example in the book. The reason why I partially note this as wrong is that even in the 70s people recognized that supersonic travel had real concrete issues with no solution in sight. I don't think LLMs share that characteristic today.

I don't see any solution to hallucinations, nor do I see any solution in sight. I think that could count as a concrete issue that would stop them.

  • Vision and everyday-physics models are the answer: hallucinations will stop when the models stop thinking in words and start thinking in physical reality.

    • The way you phrased it reveals how these model providers have framed the conversation in their favor. Models don’t “think.”

    • They had easy access to a large corpus of writing to train on, way larger than any human being trained their own language model on. I can't see where they are going to find a large corpus of physical interaction with reality to train that kind of model.

      1 reply →

Yeah, and with LLMs the thing I can't shake, however, is that this time it's pretty strongly (maybe parasitically) latched onto the aggregate progress of Moore's law. Few other technologies have enjoyed such relatively unfettered exponential improvement. It's like if skyscraper materials double in strength every n years, and their elevators approach teleportation speed, the water pumps get twice as powerful, etc., which would change the economics vs the reality that most of the physical world doesn't improve that fast.

Was the problem that supersonic flight was expensive and the amount of customers willing to pay the price was even lower than the number of customers that could even if they wanted to?

  • From what I had read in passing and remember.

      - They were loud (sonic booms were nasty).
    
      - They were expensive to maintain and operate. Guzzlers. (Britain and France clung to them as a matter of pride/ego)
    
      - They were narrow and uncomfortable. I have seen videos where there is space only for one stewardess to walk. I had been inside of one in Seattle museum. Very cramped.
    
      - As you mentioned, ticket cost was high.
    
      - I suspect people traveled in these mostly for bragging rights.

    • You made this point in passing, but it's so relevant to LLMs I wanted to highlight it: The development and operational cost was heavily subsidized by the British and French governments, because having an SST was a point of national prestige.

  • Yeah, basically. Nobody wanted to pay $12,000 to be in a plane for three hours when they could pay ~$1200 to be in one for six hours. Plus, they used up a lot of fuel. That made them real vulnerable to oil price spikes.

    Contrast that with modern widebody jets, which fly ~300 people plus paid cargo on much more fuel-efficient engines.