← Back to context

Comment by voidhorse

4 days ago

I completely agree that we don't know enough, but I suggest that that entails that the critics and those who want to be cautious are correct.

The harms engendered by underestimating LLM capabilities are largely that people won't use the LLMs.

The harms engendered by overestimating their capabilities can be as severe as psychological delusion, of which we have an increasing number of cases.

Given we don't actually have a good definition of "thinking" what tack do you consider more responsible?

> The harms engendered by underestimating LLM capabilities are largely that people won't use the LLMs.

Speculative fiction about superintelligences aside, an obvious harm to underestimating the LLM's capabilities is that we could effectively be enslaving moral agents if we fail to correctly classify them as such.

  • If the models were conscious, intelligent, suffering and could think, why wouldn't they tell us ?