← Back to context

Comment by borsch_not_soup

6 hours ago

Interesting, I’ve always thought neural network progress was primarily bottlenecked by compute.

If it turns out that LLM-like models can produce genuinely useful outputs on something as constrained as a Commodore 64—or even more convincingly, if someone manages to train a capable model within the limits of hardware from that era—it would suggest we may have left a lot of progress on the table. Not just in terms of efficiency, but in how we framed the problem space for decades.

  YOU> hey
  C64> HELLO! RE SOUNDS ME. MEFUL!

60s per token for that doesn't strike me as genuinely useful.

Very, very cool project though!

  • not useful in a disaster scenario:

    YOU> HELP I'M DROWNING

    C64> YOU' HERE!

    YOU> OH NO I'M ON FIRE

    C64> IGLAY!

    YOU> IM BEING SWALLOWED BE A SNAKE

    C64>

    YOU> BIRDS ARE NIPPING ON ME

    C64> YOU

    • Reminds me of Terry Davis' random word generator :')

      Maybe there is deeper wisdom in there that we have yet to unearth