



Study with the several resources on Docsity
Earn points by helping other students or get them with a premium plan
Prepare for your exams
Study with the several resources on Docsity
Earn points to download
Earn points by helping other students or get them with a premium plan
Community
Ask the community for help and clear up your study doubts
Discover the best universities in your country according to Docsity users
Free resources
Download our free guides on studying techniques, anxiety management strategies, and thesis advice from Docsity tutors
Alan Turing's 'imitation game' or 'Turing test', a behavioral test for machine intelligence proposed in his 1950 paper. The test involves an interrogator attempting to identify a human or machine respondent based on textual interactions. Turing predicted that by 2000, computers would be able to perform the test convincingly, but the test's validity and limitations continue to be debated.
What you will learn
Typology: Exercises
1 / 5
This page cannot be seen from the preview
Don't miss anything!
MARK BISHOP
KEYWORDS: Turing test, Imitation game
This issue of the Kybernetes journal is concerned with the philosophical question
(^1) The (British) society for the study of Artificial Intelligence and the Simulation of Behaviour. (^2) To coincide with the Loebner Prize competition held at the University of Reading (UK)
on the 12th October 2008, the AISB elected to sponsor a one day ‘invited-speaker’ symposium to present an alternative, formal, academic critique of issues around the Turing test. The day commenced with talks from three eminent speakers (Baroness Greenfield; Michael Wheeler & Selmer Bringsjord) who offered a personal context to, and their perspective on, the Turing test. These presentations were followed in the afternoon session with the invitation to four subsequent speakers (Andrew Hodges; Luciano Floridi; Margaret Boden; Owen Holland) to address specific matters related to the Turing test (e.g. definitional; adequacy; tests in other modalities; technical/computational issues). The day ended with a short round-table discussion regarding some of the issues raised during the day 1
2 MARK BISHOP
“The extent to which we regard something as behaving in an intel- ligent manner is determined as much by our own state of mind and training as by the properties of the object under consideration. If we are able to explain and predict its behaviour or if there seems to be little underlying plan, we have little temptation to imagine intel- ligence. With the same object therefore it is possible that one man would consider it as intelligent and another would not; the second man would have found out the rules of its behaviour. It is possible to do a little experiment on these lines, even at the present stage of knowledge. It is not difficult to devise a paper machine which will play a not very bad game of chess. Now get three men as subjects for the experiment A, B, and C. A and C are to be rather poor chess players, B is the operator who works the paper machine. (ln order that he should be able to work it fairly fast it is advisable that he be both mathematician and chess player.) Two rooms are used with some arrangement for communicating moves, and a game is played between C and either A or the paper machine. C may find it quite difficult to tell which he is playing. (This is a rather idealized form of an experiment I have actually done.^3 ) Subsequently, in the initial exposition of the imitation game presented in the 1950 paper [11], Turing called for a human interrogator (C) to hold a conversation with a male and female respondent (A and B) with whom the interrogator could communicate only indirectly by typewritten text. The object of this game was for the interrogator to correctly identify the gender of the players (A and B) purely as a result of such textual interactions; what makes the task non-trivial is that (a) the respondents are allowed to lie and (b) the interrogator is allowed to ask questions ranging over the whole gamut of human experience. At first glance it is perhaps mildly surprising that, even after many such textual interactions, a skilled player can determine (more accurately than by chance) the correct gender of the respondents.^4
(^3) Some commentators (e.g.Whitby [1] and Shah & Warwick (herein)) have suggested that Tur-
ing didn’t intend the imitation game to be the specification of some fully operational procedure to be performed by future machine intelligence researchers as a yardstick with which to evaluate their wares, but merely as a thought experiment, a “philosophical ice-breaker” (ibid), “attempt- ing to deal with the ill-definition .. of the question .. can machines think?” (Wiggins [15]). The fact that Turing personally enacted this first version of the imitation game offers perhaps partial evidence against this interpretation. (^4) Here Turing’s Victorian-esque parlour game describes a scenario perhaps not unfamiliar to
that many twenty-first century video gamers encounter when participating in a large multi-user virtual world - such as World of Warcraft or Second Life - where in-game avatars controlled by real-world players may often fail to reflect the gender they claim to be; the controller may be female and the avatar male or vice versa.
4 MARK BISHOP
minutes of questioning in total or did he mean five minutes of questioning per respondent? Furthermore, although Turing specifically describes playing the imitation game with ‘average’ interrogators, some commentators - perhaps remembering Kas- parov’s titanic series of games against chess playing machines - hint at a ‘strong’ version of the imitation game; where the interrogator is an expert interrogator, the game is played as an open ended conversation and the test is for full ‘human indistinguishability’ (cf. Hugh Loebner’s specification for the gold medal prize in his version of the Turing test^8 ). In 2008 the organisers of the annual Loebner bronze medal Prize elected to put Turing’s 1950 prediction to the test in the first - least demanding - manner, by enacting a set of five minute Turing tests for the bronze medal prize; specifically, each interrogator was allowed a total of five minutes to respond to both entities (the human and the computer). As a consequence the expected interaction time with the computer program was just two and a half minutes. However, in 2008 even this minimal ‘five minute’ claim proved optimistic as Elbot - evaluated as the best computer program in this competition - achieved a maximum deception rate of 25% over two and a half minutes of interaction; still 5% short of the 30% deception rate Turing had predicted in 1950. Nonetheless it seems very likely that in the next few years Turing’s predictions for a ‘time limited’ Turing test will be met; whether that means at that juncture “general educated opinion will have altered so much that one will be able to speak of machines thinking without expecting to be contradicted” (as Turing asserted) is very doubtful as, in the 50 plus years since the paper was first published, the status of the Turing test as a definitive measure of machine intelligence and understanding has been extensively critiqued^9. In 2008 the AISB sponsored an invited speaker symposium on the Turing test at the University of Reading in the hope of eliciting further clarity in the inter- pretation of the test, further insight into its implications and further reflection
(^8) In 1990 Hugh Loebner agreed with ‘The Cambridge Center for Behavioural Studies’ to un-
derwrite a contest designed to implement a Turing-style test. Dr. Loebner pledged (i) a ‘Grand Prize’ of $100,000 and a ‘solid 18 carat gold medal’ for the first computer program whose re- sponses were indistinguishable from a human’s and (ii) an annual prize - $3,000 in 2010 - and bronze medal to be awarded to the most human-like computer program (i.e. the best entry relative to other entries that year, irrespective of how good it is in an absolute sense). A com- prehensive description of the 2008 competition by Huma Shah and Kevin Warwick is presented herein. (^9) Perhaps the best known criticism of ‘a Turing style test of machine understanding’ comes
from John Searle. In the Chinese Room Argument (CRA) [14] Searle endeavours to show that even if a computer behaved in a manner fully indistinguishable from a human (when answering questions about a simple story) it cannot be said to genuinely understand its responses and hence the computer cannot be said to genuinely think (for recent discussion of the CRA see Preston & Bishop [13]).
THE IMITATION GAME 5
as to its status as a [practical] measure of machine intelligence^10. However the breadth and depth of the material presented in this special issue clearly illustrate that Turing’s imitation game continues to present novel insights into mind and machine and any hope for a ‘near final’ word on the imitation game remains as far off as ever. On behalf of the AISB and Kybernetes, I would like to thank all the expert contributors to both the 2008 symposium and this volume for making time to address these ever fascinating issues.
References
[1] Whitby, B. (1996) The Turing Test: AI’s Biggest Blind Alley? In: Millican, P. & Clark, A. (eds) (1996) Machines and Thought: The Legacy of Alan Turing. Mind Association Occasional Series 1, pp. 53-62. New York & Oxford: Oxford University Press. [2] Colby, K.M., Hilf, F.D., Weber, S. & Kraemer, H.C. (1972) Turing-like indistinguishability tests for the validation of a computer simulation of paranoid processes, Artificial Intelligence 3: 199– [3] Copeland, B.J. & Proudfoot, D. (2005) Turing and the Computer, in Copeland, B.J. (ed) (2005) Alan Turing’s Automatic Computing Engine, Oxford & New York: Oxford University Press. [4] Copeland, B.J. (2000), The Turing Test, Minds and Machines, 10: 519–39. [5] Copeland, B.J. (ed) (2004) The Essential Turing. Oxford: Clarendon Press: 488. [6] Genova, J. (1994) Turing’s Sexual Guessing Game, Social Epistemology, 8: 313–26. [7] Moor, J. (2001), The Status and Future of the Turing Test, Minds and Machines, 11: 77-93. [8] Piccinini, G. (2000), Turing’s Rules for the Imitation Game, Minds and Machines, 10: 573-85. [9] Traiger, S. (2000) Making the Right Identification in the Turing Test, Minds and Machines, 10: 561-572. [10] Turing, A.M. (1948) Intelligent Machinery. National Physical Laboratory Report, 1948, in Copeland, B.J. (ed) (2004) The Essential Turing, Oxford & New York: Oxford University Press. [11] Turing, A.M. (1950) Computing Machinery and Intelligence, Mind 59, pp. 433–460. [12] Turing, A.M., Braithwaite, R., Jefferson, G. & Newman, M. (1952) Can Automatic Cal- culating Machines Be Said To Think? in Copeland, B.J. (ed) (2004) The Essential Turing. Oxford: Clarendon Press: 487–506. [13] Preston, J. & Bishop, M. (eds) (2002) Views into the Chinese Room: New Essays on Searle and Artificial Intelligence, Oxford & New York: Oxford University Press. [14] Searle, J. (1981) Minds, Brains, and Programs, Behavioral and Brain Sciences, 3: 417-57. [15] Wiggins, G. (2007) Personal communication, Geraint Wiggins, Professor of Computational Creativity, Goldsmiths, University of London, UK.
(^10) The AISB will host a second Turing symposia at their spring convention 2010 to continue dis- cussion of this question http://www.cse.dmu.ac.uk/˜aayesh/TuringTestRevisited/Welcome.html