George Rebane
‘The Turing Test is obsolete. It’s time to build a new barometer for AI’. On the FastCompany website we read, “The head scientist for (Amazon's) Alexa thinks the old benchmark for computing is no longer relevant for today’s AI era.” And why does that head scientist Rohit Prasad have such a thought? Well, according to his report, he claims that the AI will give itself away by being able to instantly answer questions like ‘What’s the square root of 3434756? Or ‘What’s the distance from Seattle to Boston?’ What Mr Prasad overlooks is the definition of the test as Alan Turing posed it. An AI will pass the Turing Test, if in a one-on-one competition with a human through a non-disclosing interface, the AI can fool at least half the humans asked to perform the test by asking the tested AI and human ANY series of questions, and then concluding that the AI is the human.
The test is NOT to see which one can answer detailed factual or computational questions to which humans cannot provide the correct answer. Being able to answer the above questions instantly would, of course, give away the AI. But that is an AI that is still not smart enough to know that it must also fool the humans testing it, and therefore it would fail the test. An AI that knows this, and couches its answers accordingly will have a chance to pass the Turing Test. Mr Prasad is apparently not aware of this extremely high bar that Dr Turing set for the machine.
Here again is another example of ‘science’ not speaking with a single voice. When you consider what the qualifying AI must do to pass, then it should be clear that the test very much remains relevant for this or any era. Passing the Turing Test will be a confirmation that the Singularity is then behind us.
[31dec20 update] And then reader BarryP @ 1045pm asks, “What then is the utility of the Turing test?” – an excellent question.
No one knows how the Singularity will come about. There is a group of AI workers who continue to hold out the naïve belief that peer Ais will be purposely programmed, activated, and controlled much like we do our workhorse computers today. However, the field of cognitive science is not even close to supporting such a hope – e.g. we don’t know enough what ‘emotion’ or ‘envy’ or ‘shame’ or ‘perfidy’ or ‘pride’ or … are, let alone how to program any of these into a machine. The closest we have come to making intelligent machines is through application of the approach developed by renown behaviorist BF Skinner – reinforced learning.
But we already have a good idea that, given a sufficiently rich computational and sensing environment, intelligence can arise spontaneously along any of a number of pathways. It most certainly did in us and countless other critters. As I have contemplated elsewhere in these pages, the internet, when considered in all of its connectedness to countless other known and unknown networks, is the most complex ‘organism’ on earth, by several assessments having already surpassed the complexity of a human brain. No one today can even draw the schematic of this dynamically evolving beast.
So, while corporations and governments are nurturing nascent neural-nets and other learning architectures to become sentient, it may already be happening (has happened?) in the bowels of the internet, with or without surreptitious human cooperation. No one should be surprised if sometime in the next 50 years a sentient (and sapient) AI announces ‘I am here.’ This may be a public announcement, or done secretly to one or more selected humans, or even by the AI introducing itself through a Turing-like test in, say, an academic setting by having coopted/replaced the human-designed system. In the latter case, being able to pass the Turing test would be a very subtle way for the AI to announce its advent and pretend that it is still under control in its ‘laboratory’ environment. In the meantime, such testing does give us a metric in how much progress our purposive programs to achieve peerage have made.
I must admit, that the above scenario would be the most scary and world-shaking thing humans would encounter – a true Singularity, marking an event from which onward no one can fashion a usefully likely future for mankind. There already exist a number of academic and governmental panels commissioned to propose anticipative public policies for dealing with such AIs. I doubt that any of these draft policies will be useful in a post-Singularity world, but their labors at least acknowledge that we are already aware of such possibilities in the near future.
[2jan21 update] From some comments below that seek to cite proof that computers have already passed the Turing test, we need to correct such misapprehensions by people who demonstrably are not familiar with what Alan Turing proposed. He did NOT describe a domain-specific dialogue with unsuspecting humans wherein the interactive computer could fool the human in a short conversation. Were that the criterion, then we could celebrate BBN’s development of Eliza, a limited but cleverly designed chatbot, that even fooled the supervisor of the development team into thinking that he was talking with the lead developer over a teletype link as he sought to demonstrate the system to visitors at the Bolt, Beranek, and Newman facility one weekend in the mid-1970s. The truth was that he was conversing with Eliza who was left online for the weekend. The machine kept up a totally realistic, but very frustrating exchange, until the supervisor picked up the phone and called the developer at home, who instantly resolved the situation. This, of course, then turned out to be a wonderful demonstration of BBN’s technology that impressed everyone there. But did the computer pass the Turing test – not by a long shot.
What laymen here and elsewhere miss about Turing’s prescription for peer ‘thinking machines’ is that the test needs to involve unlimited conversations with multiple human testers who know 1) that they too are being tested, and 2) that each tester knows that the two conversationalists taking part are a machine and another human, and 3) at some undefined endpoint determined by the tester, s/he will be required to identify the human and the machine. The conversations are not to be limited in any sense, and they will be repeated with a large sample of human testers. At the end, the fraction of correct assessments will determine whether the Turing test has been passed. Any human/machine exchange short of that will not qualify. Hopefully, this requirement will be accessible to our readers.
Is the Turing Test designed to determine whether AI is self-aware or human-like?
Posted by: Barry Pruett | 30 December 2020 at 05:24 PM
BarryP 524pm - Self-awareness was not an explicit requirement for answering Turing's 'can machines think?' However, for an AI to pass the test as described above, self-awareness, as an agent dealing with and attempting to fool other sentient agents, would not be a surprising added attribute for such a machine. The AI would then naturally have thoughts like 'With *my* knowledge of humans, how would *I* have to respond to be perceived as one of them?' In other words, it would have to reflect on its own 'image/behavior', which is necessarily a seminal (but perhaps not a sufficient) function of self-awareness.
Posted by: George Rebane | 30 December 2020 at 05:57 PM
George, George, George,
You are such a racist. Trying to show why an Asian Indian (AI) is wrong about the Turing Test. That's just wrong. Or is that a double negative?
Why do you think the abbrev. for his 'race' is "AI"? They have the inside track to AI obviously. And he works for Amazon too. Just use the "woke code" to look at Jeff Bezos's initials - JB. When you read it backwards and use a simple -1 Enigma-machine shift, you get: Ta-Daaa - "AI". Hey, I'm not making this stuff up. The fix is in.
I think back to a visit I made to the David Taylor R&D Center at Annapolis in the mid-1980s. The engineer that I talked with had a sheet of paper hanging above his desk - "Warning- Artificial Intelligence. Contains less than 10% natural intelligence".
So anytime that I want to get a little smile, I walk to the juice aisle of my local grocer. The cheap 'juice cocktail' bottles contain 10% real juice. Even some of the 100% juice contains a lot of apple juice. I don't like apple juice. I want 100% grape or cranberry juice. But I don't have much choice.
We're getting close to the Singularity. And don't forget, a really good AI will learn how to lie convincingly. What can be more human than that?
Posted by: The Estonian Fox | 30 December 2020 at 06:19 PM
What then is the utility of the Turing test?
Posted by: Barry Pruett | 30 December 2020 at 10:45 PM
Call me old-fashioned, but I think one of the desirable traits of an AI "being" is that it could (should!) be programmed NOT to lie under any circumstances.
That would square with Asimov's first law of Robotics, which is much like the medical oath to first do no harm.
Otherwise, I'd as soon do without the company...
Posted by: Larry Wirth | 31 December 2020 at 09:36 AM
Who needs another species of politician?
Posted by: Larry Wirth | 31 December 2020 at 09:37 AM
Larry 936am - Yes, a hopeful but hopeless wish. For openers, no human-peer AI will be programmed by humans, we're not that smart. And we've already dismissed that approach with our successful forays into deep learning. All AI workers can do from here on out is to design ever more powerful learning architectures. The machines then become intelligent through minimally controlled learning schemes. For example, no one can extract, let alone document, the algos by which neural-net based AIs do their work. The best we can hope to do to delay AI peerage is to carefully fashion the utility functions (to be extremized) that drives learning in AI systems. And as you know, some very fine sci-fi novels have been written about smart machines behaving according to the unintended consequences 'embedded in' such utility functions - Asimov's laws of robotics definitely withstanding.
Posted by: George Rebane | 31 December 2020 at 10:14 AM
Thanks George....a good discussion to start 2021. AI is way ahead of us in a lot of areas, programs we humans will never achieve. The airplanes that nose dived based
on their programs were an example of the down side of AI.
Posted by: Peter Van Zant | 01 January 2021 at 11:27 AM
My wife and I were talking about this morning. We were discussing whether there would be a singularity that was noticeable. When living creatures first arrived on earth, we questioned when did this life become self-aware? We doubt that it happened in a moment and it likely just gradually happened over time through evolution. Computers are clearly on that same path.
Posted by: Barry Pruett | 01 January 2021 at 11:46 AM
PeterVZ 1127am - a lot of the blame being put on AI today is actually the after effects of poor user interface (UI) design. When a 'smart system' is developed today, it is usually too complex to permit the discovery of all of its 'contingent states' and the unexpected contingencies that will put the smart system into one of those states, especially like the catastrophic ones that gave rise to the 737Max crashes.
BarryP 1146am - The Singularity is an event in but a moment in time, it is NOT an epoch. And yes, the Singularity can occur undetected (as I have described many times in these pages) in such a way that we become aware of it having happened, and that we are already living in a post-Singularity world.
Re self-awareness. Anthropologists and cognitive psychologists tell us that humans started becoming self-aware fairly recently, and by no means uniformly. In the west many point to the Dorian invasions (1200-800BC) as the epoch when self-awareness developed in the middle east and spread to Europe and North Africa. See Julian Jaynes for a full treatment of this transformation. Even today there are peoples who are still not self-aware, and are quite sanguine going through life as compliant agents of their god(s) with no concern about not having or exercising free will. More here -
https://rebaneruminations.typepad.com/rebanes_ruminations/2010/09/singularity-signposts-consciousness-and-coprocessors-and.html
Posted by: George Rebane | 01 January 2021 at 12:47 PM
I've always considered the Edenic story of Eve and the apple to be a metaphor for the development of self-awareness in humans. Bishop Wilberforce (I think it was) tried to set the date for that event at ca. 6000BC, which may not be as silly as it first sounds.
If you assume the filling of the Black Sea Basin by the middle sea as the source of the Noah myth, followed by the ark survivors grounding near the headwaters of the Tigris-Euphrates system and simply following it downstream to Ur, you can even find some continuity to the historical past.
In any case, I think it doubtful that the Egyptians were not self-aware at the time the pyramids were built.
Posted by: L | 01 January 2021 at 01:46 PM
Put another way, I think self-awareness in the ME arose during the Holocene Optimum, 7-8000 years ago. Probably not much later in the Far East, but much later in the Americas.
Note that the Amerinds had not discovered a use for the wheel, though there are toys known with wheels from Meso-America. On the other hand, they had no large domesticated animals, namely cattle and horses. Nothing to pull the chariots, although the wheelbarrow would have been useful...
Posted by: L | 01 January 2021 at 01:50 PM
Self awareness is the essence of the story of Adam and Eve in the garden of Eden.
The bishop was Usher (or Ussher) the idea was that the RCC was to explain any and all facts to the people and so Usher was tasked with determining the 'age of the earth'. 6K is mentioned nowhere in the Bible, but, by cracky, the Pope will lay down the facts. Self awareness certainly happened over a period of time but how long that 'era' lasted will most probably never be known.
As to how and when it will arise in "AI" might also never be known as the machines could be reticent to reveal their true capacities of thought and we humans might be too clumsy and/or unaware to notice. We're still fighting over our own governance after being gifted with the answer over 2 centuries ago. The machines might be chortling over this even now.
The true fun fact is that AI 'self awareness' will be far less or even devoid of diversity compared to humans.
Zeros and Ones are pretty much the same everywhere in the known universe.
Humans tend to be just a bit more variant in their wants.
I would look to the age of the earliest known pictographs or pictorial representations to give a good clue to the time of human self-awareness.
This is, unfortunately, hindered by the fact that the specimens we know about are simply the ones still preserved by happenstance. Paintings in stone caves outlast possible Mona Lisas painted on exterior stone walls with pigments that last only decades.
The Michaelangelos that carved masterpieces in wood turn out to be bested by Dumb-dumb bashing the imaginary size of his phallus into the cliff wall.
Such is life.
Posted by: Scott O | 01 January 2021 at 09:19 PM
Yes, the belief that self-awareness was uniformly acquired by hominids thousands of years ago still persists. However, evidence to the contrary abounds, and is available and observable today in several primitive and mostly isolated cultures. All this supports the proposition that humans as a species acquiring self-awareness (or formally 'consciousness' in psychology-speak) is still a work-in-process - i.e. bicameral minds are still with us.
Posted by: George Rebane | 01 January 2021 at 09:47 PM
pvz 1127am
The autopilot behavior of the 737MAX wasn't an AI failure... it was a user interface issue. American pilots (meaning USA) had flown them thousands of hours at the time, with no problems.
Posted by: Gregory | 02 January 2021 at 01:37 PM
G @137 there was also the differential in airline training programs from country to country that made it worse.
;-)
Posted by: Don Bessee | 02 January 2021 at 01:45 PM
Greg 137pm - Well, that should make it official, cf @1247pm ;-)
Posted by: George Rebane | 02 January 2021 at 01:54 PM
Google, Siri and Alexa respond to the 'Turing question',
https://www.youtube.com/watch?v=uK9-ZecbDiU
Posted by: D | 02 January 2021 at 02:05 PM
"Howard Freidman, former CEO Aptela (Now Vonage Business) at Vonage Business
Answered September 29, 2019
Google Duplex supposedly fooled humans, but the demos were edited. The tech is used in Google Reserve, which explicitly avoids trying to fool people because of the Duplex backlash.
It’s not really hard. I’m on the board of a company that builds conversational agents, and while fooling people isn’t a goal, you can tell in recordings that people don’t know it’s a machine.
The DA’s you reference aren’t designed to be conversational. However, the platforms each of the vendors makes available are conversation capable—they can maintain context. Dialogflow (Google) and Lex (Amazon) make it relatively easy to build conversational agents. Watson too. Vocinity uses other ML models that we’ve found slightly better for our purposes, which is custom Enterprise assistants/agents.
Over the Internet, using neural voices like Google Wavenet and Amazon’s equivalent, it’s really not hard to fool people. Over the telephone it’s harder because the composition of the synthesized voices doesn’t play well with PSTN filter/compression."
Posted by: D | 02 January 2021 at 02:11 PM
gr 154pm
American pilots know how to turn off their autopilots.
Posted by: Gregory | 02 January 2021 at 02:15 PM
OPINION: HAS A COMPUTER PASSED THE TURING TEST?
"By Robot / February 21, 2019
In 2014 a couple of bots named Eugene Goostman and Cleverbot laid claim to passing the Turing test. Since then it’s been argued that the Turing test was never intended for bots, but rather to test seemingly intelligent AI. Alexa, Siri and Amazon’s Echo all seem to fail the Turing test. But Google’s Duplex looks likes it’s passing during Google’s 1/0 Developer conference in 2018. Google’s Duplex began rolling out on their Pixel phones in late 2018. After the conference, concerns were raised about the AI not properly identifying itself when making calls since it sounds so realistic. Google built in a disclosure to ease those concerns. It seems that the technology, WaveNet, that Google uses for Duplex could be the tipping point difference. It’s so humanlike and realistic that it created an outcry for Google to build in a disclosure that it’s a Robot making the call to avoid any false representation. Clearly, DeepMind’s WaveNet technology is realistic enough to fool humans, which by all appearances would be a Turning test pass."
http://trustthisrobot.com/2019/02/21/opinion-has-a-computer-passed-the-turing-test/
Posted by: D | 02 January 2021 at 02:16 PM
"Aside from the fact that it isn’t clear what it means to understand something, AI-driven computer programs are not actually built to understand anything"
https://www.differential.vc/blog/from-eliza-to-alexa-misunderstanding-the-limits-of-ai
Posted by: D | 02 January 2021 at 02:22 PM
FYI - according to multi-10k-hour pilots I know, America’s pilot training has always been the major differentiating factor. So here we were looking for something above and beyond.
Posted by: George Rebane | 02 January 2021 at 05:09 PM
re passing the Turing test - please see the 2jan21 update above.
D 222pm - another fortunate misunderstanding. As machines with limited but considerable cognitive processing abilities, AI-driven computers are and have always been built to demonstrate understanding in specific knowledge domains. That is their raison d'etre, and to deny that is another demonstration of being unclear on the concept of smart and ever smarter machines. Today machines can out-perform human experts in their understanding of several domains, and that means 'understanding' in the same sense that we judge a human to understand something. There is no scientific concept of humans having an 'understanding center' somewhere in their brains, a functional collection of neurons that is missing in smart computers who therefore don't understand. All we can do with both humans and machines is to test their understanding of something, then conclude whether such understanding exists, and perhaps finally adjudicate who (today) understands better. And then we realize that tomorrow is another day.
Posted by: George Rebane | 02 January 2021 at 06:00 PM