Comment by _russross

17 hours ago

Turing himself argued that trying to measure if a computer is intelligent is a fool's errand because it is so difficult to pin down definitions. He proposed what we call the "Turing test" as a knowable, measurable alternative. The first paragraph of his paper reads:

> I propose to consider the question, "Can machines think?" This should begin > with definitions of the meaning of the terms "machine" and "think." The > definitions might be framed so as to reflect so far as possible the normal use > of the words, but this attitude is dangerous, If the meaning of the words > "machine" and "think" are to be found by examining how they are commonly used > it is difficult to escape the conclusion that the meaning and the answer to the > question, "Can machines think?" is to be sought in a statistical survey such as > a Gallup poll. But this is absurd. Instead of attempting such a definition I > shall replace the question by another, which is closely related to it and is > expressed in relatively unambiguous words.

Many people who want to argue about AGI and its relation to the Turing test would do well to read Turing's own arguments.

The Turing test ended up being kind of a flop. We basically passed it and nobody cared. That's because the turing test is about whether a machine can fool a human, not about its intelligent capabilities per se.

  • No, it's because certain people moved the goal posts. Nothing an LLM does or will do will make them belive that it's "intelligent" because they have a mental model of "intelligence" that is more religious than empirical.

    • We don’t have agents that are able to work entirely autonomously, even in the coding realm, which is where they seem to be most valuable. In fact, they’re seemingly not even close to replacing software engineers.

      1 reply →