intelligent humans who appear less than bright
whereas there are some machines that clearly appear
smart.
It can be seen from the examples given that some
judges could be more susceptible to deception.
Others have a biased perspective on ‘humanlike
conversation’. This may have led judges to
misclassify hidden interlocutors, even though they
were tasked with initiating conversations. Judges
were given the possibility of asking or discussing
whatever they wanted: the conversations were
‘unrestricted’. The ‘hidden humans’ were asked not
to behave like machines and to protect their identity.
However, each hidden human interpreted that
instruction to ‘foils for the machines’ in their own
humanlike way.
Not all the invited machines were designed to
imitate humans. Elbot, for example, from Artificial
Solutions has a robot personality. However, all are
designed to mimic human conversation and avoid
correctly answering mathematical questions, as
Turing had suggested. Essentially the machines are
merely trying to respond in the sort of way that a
human might.
Whatever the standing of the Turing test in the
reader’s mind, what we hope is evident from the
transcripts presented in this paper is that it is
certainly not a trivial, simple exercise. Indeed it is a
surprising indication of how humans communicate
and how the human judges might be easily fooled
based on their assumptions and individual ideas
about intelligence. Insights can lead to improved
design of intelligent agents, to make their
conversation more humanlike and build trust
between the natural and the artificial conversation
agent.
4 CONCLUSIONS
How humans talk in stranger to stranger situations
suggest general techniques for successful human-
intelligent agent interaction, in e-commerce for
example. We suggest that intelligent agents ask
more, not just to improve their conversational
ability, but to understand the human user. We
recommend that developers
a) Do not assume knowledge held by human
interlocutors
b) Appreciate that humans cannot always formulate
their enquiry clearly
c) Develop the Intelligent Agent to probe further
and ask more questions encouraging human
interlocutors to clarify their needs
d) Be prepared for mischievous users who will lie
to confuse the intelligent agent.
Lastly, the authors are continuing their Turing test
work, following the third event at The Royal Society
London in June 2014. The results from that
experiment are being analysed and will be sent for
peer review.
ACKNOWLEDGEMENTS
Harjit Mehroke for Figure 1. Bletchley Park, UK,
the judges and hidden humans, and the developers of
the machines that took part in the 2012 experiment.
REFERENCES
Fowler, H., and Fowler, F. (Eds.). (1995). The Concise
Oxford Dictionary of Current English (9th ed., p.
486). Oxford: Clarendon Press.
Hayes, P. and Ford, K. 1995. Turing Test Considered
Harmful. Proceedings of the Fourteenth International
Joint Conference on Artificial Intelligence. Vol. 1.
Montreal, August 20-25: pp. 972-7.
Hodges, A. 1992. Alan Turing: the Enigma. Vintage
Books, London.
Minsky, M. 2013. Singularity 1 on 1: The Turing test is a
joke! Youtube video interview by Nikola Danaylov:
https://www.youtube.com/watch?v=3PdxQbOvAlI.
Penrose, R. (1994). Shadows of the Mind: A Search for the
Missing Science of Consciousness. Oxford University
Press, UK.
Searle, J.R. 1997. The Mystery of Consciousness. The
New York Review of Books. NY, US.
Shah, H. 2013. Conversation, Deception and Intelligence:
Turing’s Question-Answer Game. In S.B. Cooper and
J van Leeuwen (Eds) Alan Turing: his life and impact.
Elsevier: Oxford, UK: pp. 614-620.
Shah, H., Warwick, K., Bland, I.M., Chapman, C.D. and
Allen, M. 2012. Turing’s Imitation Game: Role of
Error-making in Intelligent Thought. Turing in
Context II, Brussels, 10 October.
Shah, H. 2011. Turing’s Misunderstood Imitation Game
and IBM’s Watson Success. Keynote in 2
nd
Towards a
Comprehensive Intelligence test (TCIT) symposium at
AISB 2011, University of York, 5 April.
Shah, H., and Henry, O. 2005. Confederate Effect in
human-machine textual interaction. 5
th
WSEAS
Information Science and Communications (ISCA)
conference, Cancun, Mexico, May 11-14.
Shah, H., and Pavlika, V. 2005. Text-based Dialogical E-
Query Systems: Gimmick or Convenience?. In
Proceedings of 10th International Conference on
Speech and Computers (SPECOM), University of
IntelligentAgents-ConversationsfromHuman-agentImitationGames
267