Skip to main content

Advertisement

Log in

Imitation Game: Threshold or Watershed?

  • General Article
  • Published:
Minds and Machines Aims and scope Submit manuscript

Abstract

Showing remarkable insight into the relationship between language and thought, Alan Turing in 1950 proposed the Imitation Game as a proxy for the question “Can machines think?” and its meaning and practicality have been debated hotly ever since. The Imitation Game has come under criticism within the Computer Science and Artificial Intelligence communities with leading scientists proposing alternatives, revisions, or even that the Game be abandoned entirely. Yet Turing’s imagined conversational fragments between human and machine are rich with complex instances of inference of implied information, reasoning from generalizations, and meta-reasoning, challenges AI practitioners have wrestled with since at least 1980 and continue to study. We argue that the very fact the Imitation Game is so difficult may be the very reason it shouldn’t be changed or abandoned. The semi-decidability of the game at this point hints at the possibility of a hard limit to the powers of technology.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  • Allen, J., & Perrault, C. R. (1980). Analyzing intention in utterances. Artificial Intelligence, 15, 143–178.

    Article  Google Scholar 

  • Brooks, R. (1990). Elephants don’t play chess. Robotics and Autonomous SYSTEMS, 6, 3–15.

    Article  Google Scholar 

  • Calude, C. S. (2018). A probabilistic anytime algorithm for the halting problem. Computability, 7, 259–271.

    Article  MathSciNet  Google Scholar 

  • Cangelosi, A., Bongard, J., Fischer, M., &, Nolfi, S. (2015). Embodied Intelligence. In Springer Handbook of Computational Intelligence, Kacprzyk, J., and Pedrycz, J., eds, (pp. 697–714).

  • Dennett, D. (1991). Consciousness Explained. Boston: Little, Brown and Co.

    Google Scholar 

  • Feldman, J., & Sproull, R. (1977). Decision theory and artificial intelligence IIL the hungry monkey. Cognitive Science, 1, 158–172.

    Article  Google Scholar 

  • Ford, K.M., &, Hayes, P. (1995). Turing Test Considered Harmful. In Proceedings of IJCAI 1995, pp. 972–977

  • Ford, K., Hayes, P., Glymour, C., & Allen, J. F. (2016). Cognitive orthoses: Toward human-centred AI. AI Magazine, 36(4), 5–8.

    Article  Google Scholar 

  • French, R. (1990). Subcognition and the limits of the turing test. Mind, 99(393), 53–65.

    Article  MathSciNet  Google Scholar 

  • French, R. (2012). Dusting off the turing test. Science, 336(6078), 164–165.

    Article  Google Scholar 

  • Good, I. J., & Turing, A. (1953). The population frequencies of species and the estimation of population parameters. Biometrika, 40(3–4), 237–264.

    Article  MathSciNet  Google Scholar 

  • Harnad, S. (2001). Minds, machines, and Turing: the Indistinguishability of Indistinguishables. Journal of Logic, Language, and Information, 9(4), 425–445.

    Article  Google Scholar 

  • Horn, B. (1985). Computer Vision. Cambridge: MIT Press.

    Google Scholar 

  • Kelly, S. (2017). Endurance: A year in space, a lifetime of discovery, Decle Edge.

  • Kleene, S. C. (1952). Introduction to metamathematics. Amsterdam: North-Holland.

    MATH  Google Scholar 

  • Kyburg, H. E., Jr. (1974). The logical foundations of statistical inference (Vol. 65). Berlin: Springer Science & Business Media.

    Book  Google Scholar 

  • Kyburg, H. E., Jr. (1983). The reference class. Philosophy of Science, 50(3), 374–397.

    Article  MathSciNet  Google Scholar 

  • Lanier, J. (2017). Dawn of the new everything: Encounters with reality and virtual reality. New York: Henry Holt and Company.

    Google Scholar 

  • Levesque, H.J. (2011). The Winograd schema challenge. In Logical Formalizations of Commonsense Reasoning: Papers from the 2011 AAAI Spring Symposium. Technical Report SS-11–06. AAAI Press, Palo Alto

  • Levesque, H.J. (2009). Is it Enough to get the Behavior Right? Proceedings of IJCAI-09 1439:1444.

  • Levesque, H. J. (2014). On our best behaviour. Artificial Intelligence, 212, 27–35.

    Article  MathSciNet  Google Scholar 

  • Levesque, H. J. (2017). Common sense, the turing test, and the quest for real AI: Reflections on natural and artificial intelligence. Cambridge: MIT Press.

    Book  Google Scholar 

  • Leviathan, Y., & Matias, Y. (2018). Google duplex: An AI system for accomplishing real-world tasks over the phone. Google AI Blog. Retrieved March 11, 2019 from, https://ai.googleblog.com/2018/05/duplex-aisystem-for-natural-conversation.html.

  • Loui, R. P. (1987). Defeat among arguments: A system of defeasible Inference. Computational Intelligence, 3, 100–106.

    Article  Google Scholar 

  • Marcus, G. (2013). Why can’t my computer understand me? The New Yorker, https://www.newyorker.com/tech/annals-of-technology/why-cant-my-computer-understand-me .

  • McCarthy, J. (1986). Applications of circumscription to formalizing common-sense knowledge. Artificial Intelligence, 28(1), 89–116.

    Article  MathSciNet  Google Scholar 

  • McCarthy, J. & Hayes, P.J. (1969) Some Philosophical Problems from the Standpoint of Artificial Intelligence. Machine Intelligence 4, (eds Meltzer, B. and Michie, D.). Edinburgh: Edinburgh University Press 463–502.

  • McDermott, D., & Doyle, J. (1980). Non-monotonic logic. Artificial Intelligence, 13(1–2), 41–72.

    Article  MathSciNet  Google Scholar 

  • Mori, M. (2012) The Uncanny Valley: The Original Essay by Masahiro Mori, translated by MacDorman, K.F., & Kageki, N. (2012). IEEE Spectrum ,pp. 1–8.

  • Neufeld, E., & Finnestad, S. (2016a). Artificial intelligence testing. In Proceedings of the Twenty-Ninth International FLAIRS Conference, pp. 158–161

  • Neufeld, E., & Finnestad, S. (2016b). The mismeasure of machines. In Proceedings of the 29th Canadian Conference on Artificial Intelligence, pp. 58–63

  • Neufeld, E., & Finnestad, S. (2016c). The Post-Modern Homunculus. In Proceedings of the European Conference on Artificial Intelligence, pp. 1670–1671

  • Neufeld, E., & Finnestad, S. (2020). In defense of the Turing test. AI & Society, to appear.

  • Neufeld, E., & Goodwin, S. (1998). The 6–49 Lottery Paradox. Computational Intelligence, 14(3), 273–286.

    Article  MathSciNet  Google Scholar 

  • Núñez Siri, J., Neufeld, E., Parkin, I., & Sharpe, A. (2020). Using Simulated Annealing to Declutter Genome Visualizations. In Proceedings of the Thirty-Third International FLAIRS Conference, pp. 201–204

  • Poole, D. (1985). On the Comparison of Theories: Preferring the Most Specific Explanation, in Procedings of IJCAI 1985, Morgan Kaufmann, pp. 144–147.

  • Poole, D. (1989). What the lottery paradox tells us about default reasoning (extended abstract). In Proceedings of KR-89, Toronto, pp. 333–340.

  • Russell, S. (2019). Human Compatible: Artificial Intelligence and the Problem of Control, Allen Laine.

  • Russell, S. & Norvig, P. (2009) Artificial Intelligence: A Modern Approach (3rd edition) Pearson.

  • Searle, J. (1980). Minds, brains, and programs. Brain and Behavioural Sciences, 3, 417–457.

    Article  Google Scholar 

  • Shannon, C. & McCarthy, J. (1956). Automata Studies. Princeton University Press, p.vi

  • Shotter, J. (2019). Why being dialogical must come before being logical: the need for a hermeneutical–dialogical approach to robotic activities. AI & SOCIETY, 34(1), 29–35.

    Article  Google Scholar 

  • Trausan-Matu, S. (2019). Is it possible to grow an I-Thou relation with an artificial agent? A dialogistic perspective. AI & SOCIETY, 34(1), 9–17.

    Article  Google Scholar 

  • Turing, A. (1936). On computable numbers, with an application to the Entscheidungsproblem. Proceedings of the London Mathematical Society, 2(42), 230–265.

    MathSciNet  MATH  Google Scholar 

  • Turing, A. (1950). Computing machinery and intelligence, Mind 433–460.

  • Reiter, R. (1980). A logic for default reasoning. Artificial Intelligence, 13(1–2), 81–132.

    Article  MathSciNet  Google Scholar 

  • The Guardian. (2014). Computer simulating 13-year-old boy becomes first the pass Turing test, https://www.theguardian.com/technology/2014/jun/08/super-computer-simulates-13-year-old-boy-passes-turing-test.

  • Van Arragon, P. (1991). Modeling default reasoning using default. User Modelling and User Adapted Interaction, 1, 259–288.

    Article  Google Scholar 

  • Wired Magazine. (2011). Spide rSpins Zero-Gravity Web in Space, https://www.wired.com/2011/06/space-spiders-action/.

Download references

Acknowledgements

As with the knowledge we expect Turing machines to master, many of the statements herein may raise questions about exceptions and edge cases. We found it necessary, for example, to not present Kyburg’s entire opus, which he spent a lifetime improving. Here, Kyburg’s formalism provides a conceptual framework for understanding the nature of the problem, and we have elaborated sufficiently to address the examples presented in the main body. “Phantom 309” is a Red Sovine tune about a ghost truck, the driver of which sacrificed his life to save a bus full of children. The Phantom 309 still haunts the west coast, picking up the occasional hitchhiker and giving him a little change for a coffee. Thanks to Rosemary Nixon for a careful edit, Braden Dubois for several reads and re-reads, the reviewers of this paper for their comments, and thanks to the many persons we have discussed this work with over the years. Thanks also to the University of Saskatchewan for providing funding for this research.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Eric Neufeld.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendices

Appendix

Appendix 1: Representing Generalizations with First Order Logic

Summing up the problems encountered during a decade of research necessarily requires oversimplification. We begin with a classic example:

Chilly-Willy is a penguin.

Donald is a duck.

Penguins are birds.

Ducks are birds.

Birds fly.

Penguins don’t fly.

We leave it to the reader to observe that by combining different subsets of this knowledge base, we can show that Chilly-Willy flies, and that Chilly-Willy does not fly. We can eliminate one of these conclusions by designating “birds fly” as a generalization, and adding a rule that an instance of a generalization can only be applied if the set of all sentences used cannot be made to generate a contradiction. We can still use an instance of the generalization to conclude Donald is a bird and can fly. These sentences allow other interesting inferences. For example, taking contrapositive forms, we conclude that if it flies, it’s not a penguin, and if it’s not a bird, it’s not a duck. In this setting, the contrapositive forms make sense, but that isn’t always the case.

Now suppose, ducks are different from most birds in that they have webbed feet. To fully incorporate this into the database, we add the following:

Chilly-Willy is a penguin.

Donald is a duck.

Penguins are birds.

Ducks are birds.

Birds fly.

Penguins don’t fly.

Birds don’t have webbed feet.

Ducks have webbed feet.

Again, this seems reasonable. But suppose we add one more sentence.

The only birds are ducks and penguins.

Although it seems unreasonable to say every bird is either a duck or a penguin, this gives a compact counterexample. Let’s explore the problem with the compact counterexample, then generalize to something more reasonable.

The counterexample goes as follows. Let Foghorn be a bird. Birds typically fly, so if Foghorn flies, Foghorn can’t be a penguin. Because everything is either a penguin or a duck, Foghorn is a duck and has webbed feet. Using the same trick, we can put together a different set of sentences and conclude that Foghorn is a penguin and can’t fly.

Here is a fuller counterexample. Let there be 1000 kinds of birds. Each kind is different from a typical bird in some way, but otherwise is a normal bird. (If the kind has no distinguishing feature from other kinds, how can it be a kind?) Plus, we have the clause that every bird must be one of the 1000 kinds. Using 999 of the generalizations contraposed, we can rule out that Foghorn is not any of those 999 kinds, and therefore must be the remaining kind and be unique in the way the remaining kind is unique.

Some readers will see this as a variation on Kyburg’s lottery paradox, others as a variation on Simpson’s paradox. Either way, the simple reasoning pattern initially proposed has collapsed. This result is our interpretation of (Poole, 1989).

Appendix 2: What Practical Certainty Buys Us

The idea of practical certainty lets us hold as beliefs a set of sentences, which written as a conjunction would contradict some fact. The classic example is a lottery, where the purchaser buys a numbered ticket, and only one number wins, as opposed to modern lotteries where the purchaser can choose their numbers. It is reasonable in such a lottery to believe, for each ticket, that it will lose. But it is not reasonable to believe that no ticket will win, since by construction a winner is drawn. (For an argument that this still holds for the modern lottery, see (Neufeld and Goodwin 1998)).

To keep the calculations simple, let’s suppose there are 20 unique tickets in the lottery. This means that the probability any ticket will lose is 0.95. Thus, it is practically certain that each ticket loses. Suppose an individual buys two tickets. The probability that both tickets lose is 0.9 – this is not a practical certainty, but a probability. In this situation, we cannot combine two practical certainties into a conjunction that is a practical certainty.

As the number of tickets gets large, one can be practically certain that if two tickets are purchased, both will lose.

Applying this to the previous ‘bird’ example, we can’t treat the conjunction that Foghorn is not any one of 999 kinds of birds as a practical certainty. This prevents the collapse of the formalism, but also limits the formalism’s inferential power.

We remark that the example above assumed buying tickets without replacement, which simplified the calculation. More generally, let A and B be any two events of probability 0.95. Thus each is a practical certainty. Using the basic identity.

P(A&B) = P(A) + P(B) – P(A or B).

(where P is probability) the probability of the conjunction could be less than 0.9 because P(A or B) might be 1.

However, if A and B are two arbitrary events of probability 0.99, we can show the lowest value of their conjunction is 0.98 using the same formula, even if the probability of the disjunction is unity, and the conjunction is a practical certainty.

Finally, we remark that the theory of epistemological probability has many nuances. A reader of an earlier draft of this paper asked the following. If 1% of ticks carry Lyme disease, then 99% of ticks do not, and thus it is practically certain that ticks do not carry Lyme disease. This is a knowledge engineering problem worth delving into.

We will use natural language representations of the knowledge rather than introduce a new formalism. To begin with, suppose a data collector has written “Of 300 ticks examined near Gormley Wood, 3 carried Lyme disease”. This might be translated to “The probability of any particular tick carrying Lyme disease is between 0.009 and 0.011”, the interval accounting for all manner of uncertainty about how the data was collected. Next we learn, “Alice Butterwick noticed a tick on her dog Mollie after a walk through Avon Gorge.” From the statistical data, Alice can infer that “the probability the tick on Mollie has Lyme disease is about 1%” (this is lifting) and therefore be practically certain that that tick does not carry the disease. If three hundred dog-owners visit the Gorge every day, it is practically certain someone’s pet will pick up a tick carrying Lyme disease. Similarly, if Alice visits the Gorge with Mollie three hundred times, Mollie is likewise certain to be exposed to a disease bearing tick.

Alice might feel differently about a beloved let getting Lyme disease than about the probability her car will start; in that case she may wish to adjust her level of practical certainty. This also brings in the complications of decision theory. If one thinks in terms of mundane lotteries (where neither positive nor negative outcomes have drastic consequences) rather than diseases, or the commonplace assumptions one makes going about daily business, the conclusions reflect common sense.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Neufeld, E., Finnestad, S. Imitation Game: Threshold or Watershed?. Minds & Machines 30, 637–657 (2020). https://doi.org/10.1007/s11023-020-09544-5

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11023-020-09544-5

Keywords

Navigation