← Back to context

Comment by mitchell_h

8 hours ago

I watched some explain how deepseak got good and the Chinese approach to LLM training. Really wish I could remember it. The premise was China thinks of LLMs not as a thing separate from hardware, but gains efficiencies at each layer of the stack. From Chips to software, it's all integrated and purpose built for training.

Wonder if Anthropic is making a mistake by focusing on "consumer" hardware, and not going super specialized.

So you watched some random video from some random YouTuber, didn't even remember who made it, so much so you didn't even remember that deepseek isn't spelled "deapseak", didn't bother to even find it or verify, and then you go asserting your memory as fact on a serious discussion forum.

Comments like yours add nothing to the discussion.

  • I belive he does have a valid point.

    You can throw money and hardware at a problem, but then someone may come along with a great idea and leapfrog you.

    Just consider that all major AI providers now use deepseeks ideas for efficient training from that first paper.

DeepSeek uses merchant silicon like everyone else.

edit: I misunderstood, I thought you were implying they designed their own GPUs. nevermind

> I watched some explain how deepseak got good and the Chinese approach to LLM training.

I distinctly remember reading a big pantie twisting from Sam Altman and Co that Chinese took their stuff, the stuff OpenAI and Co spent billions to create, and used that as the base for $0.00

It’s fake news predicated on China not being able to get GPUs. But it turns out everyone was getting them their GPUs by serial number swaps in warehouse.