Comment by armchairhacker

1 day ago

There are more tells. Rule of three, short cliche sentences.

> We know how frustrating this is, and we hope you'll give us another look once we have something to show, we’ll save your usernames!

I think it's partly human. But ex:

> Network effects aren't just a moat, they're a wall.

isn't a natural sentence.

So no evidence at all, and just your need to point out possible LLM where ever you imagine it. You could be an LLM agent.

I think you're spot on. It feels like parts were edited with AI and parts were left alone.

> This isn't just a Digg problem. It's an internet problem. But it hit us harder because trust is the product.

The statement this is making is presumably the crux of the problem (Digg cannot survive without trust!) but it's worded so poorly that it's hard to imagine someone sat down and figured these three sentences were the best way to make the point.

How is that not a natural sentence? I think people are reading into stuff. That's just good writing.

Could it be generated? Sure. But there aren't the obvious tells you act like there are.

  • Here's the context:

    "We underestimated the gravitational pull of existing platforms. Network effects aren't just a moat, they're a wall."

    It's a mixed metaphor which doesn't make any sense. There are really very few ways in which this can be considered good writing - I guess the grammar is ok even if it is nonsense.

    So let's break it down - underestimated the gravitational effects - ok, this is nice, like where it's going talking about these big competitors sucking in users, but then we have the metaphor extended to breaking point:

    Network effects are a moat, but not just a moat, they're a wall (which is really not anything like a moat). So which of these 3 things are they, and why are we mixing the metaphors of gravity (pulling in customers), moats (competitive moat) and walls (walled gardens).

    It's just all a bit nonsensical and the kind of fuzzy prose that seems superficially impressive without actually saying anything meaningful in which LLMs excel. Go try generating an article from just the heads in this article, and see how similarly it reads.

    • If you want your gradation to work, the items need to be similar and progressively stronger. That's why it doesn't work. A wall is not "stronger" than a moat. "Not a fence, a rampart" would work.

      Compare to the canonical example from Cyrano de Bergerac: ''Tis a rock! ... a peak! ... a cape! -- A cape, forsooth! 'Tis a peninsular!'

      1 reply →

    • That’s the entire point - network effects are commonly discussed as being a moat (people can’t cross without difficulty) but are actually a wall - people can’t cross and can’t view the other side. Seems simple and straightforward to me.

    • Isnt a moat and a wall pretty similar in function? They both keep people in or out of an area.

      Also werent all "moats" commonly paired with a wall in real life? As in a moat around a castle wall?

      1 reply →

  • "Network effects aren't just a moat, they're a wall." is a VERY ChatGPT way to write. It's not proof, but the parent is right that this smells a bit of AI writing.

    • It's also a VERY HUMAN way to write.

      I don't care so much about Digg, but the endless "haha, I caught you!" comments annoy me more than the rare actual AI-written content they label.

      14 replies →

The rule of three is a basic writing structure taught to 12 year olds. I know people have given up on even the basics (capitalisation) in recent years but let's not just banish structured writing to "AI".