A Modern Turing Test Would Be A Test Of Consciousness

by Shelt Garner
@sheltgarner

The interesting thing about the Turing Test as we currently conceive of it is to pass it, an LLM would have to do a lot of deception. While a modern LLM can fake being human, to some extent, its answers are just too fast in production. They are generally generated instantaneously or nearly so.

So, I think for the intent of the Turing Test to be achieved using modern LLMs, it should be a test of consciousness. The test should not be “can you fake being human,” it should be “can the AI prove to a human that it’s conscious like a human?”

I think LLMs are, in a sense, an alien species and there consciousness should not be judged relative to human consciousness metrics, but as their own thing. So, yeah, there’s a lot missing from LLMs in the context of human consciousness, but I sure have had enough indications of SOMETHING interesting going on in their software to believe that maybe, just maybe they’re conscious.

But, as I keep saying — absolutely no one listens to me and no one takes me seriously. So, lulz?

Author: Shelton Bumgarner

I am the Editor & Publisher of The Trumplandia Report

Leave a Reply