Skip to main content

Advertisement

Log in

Dreyfus on the “Fringe”: information processing, intelligent activity, and the future of thinking machines

  • Original Article
  • Published:
AI & SOCIETY Aims and scope Submit manuscript

Abstract

From his preliminary analysis in 1965, Hubert Dreyfus projected a future much different than those with which his contemporaries were practically concerned, tempering their optimism in realizing something like human intelligence through conventional methods. At that time, he advised that there was nothing “directly” to be done toward machines with human-like intelligence, and that practical research should aim at a symbiosis between human beings and computers with computers doing what they do best, processing discrete symbols in formally structured problem domains. Fast-forward five decades, and his emphasis on the difference between two essential modes of processing, the unconscious yet purposeful mode fundamental to situated human cognition, and the “minded” sense of conscious processing characterizing symbolic reasoning that seems to lend itself to explicit programming, continues into the famous Dreyfus–McDowell debate. The present memorial reviews Dreyfus’ early projections, asking if the fears that punctuate current popular commentary on AI are warranted, and in light of these if he would deliver similar practical advice to researchers today.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

Notes

  1. For a more general review of Dreyfus’ work in philosophy of artificial intelligence with some attention, for example, to the issue of embodiment with which the present paper concludes, see Brey 2001.

  2. “It’s like claiming that the first monkey that climbed a tree was making progress towards flight to the moon.” (cf. Dreyfus 2007, p. 1142).

  3. Dreyfus seemed most disappointed that “neither Dennett nor anyone connected with the project has published an account of the failure and asked what mistaken assumptions underlay their absurd optimism” with Dennett pinning the stall-out on a “lack of graduate students” (2007, p. 1141).

  4. Of the four, he holds pattern recognition to be most fundamental, noting that “resolution of the difficulties which have arrested development” in other fields “all presuppose success” in this one (1965, p. 14).

  5. Later, he bundles these three forms of information processing into what he calls “perspicuous grouping”, the ability to perceive connections without explicitly counting them out, and to recognize individual examples as “typical” of a class without explicit comparison of every other given example (1965, p. 37–46).

  6. As the chess player is “zeroing in” on opportune movements, he is effectively “unconsciously counting out” the various alternatives (1965, p. 53).

  7. Structuring a problem “does not involve merely the given parts and their transformations” but “works in conjunction with the material that is structurally relevant but is selected from past experience.” (Dreyfus 1965, p. 29) This insight bears on prospective agency through backpropagation of perceived error as realized in contemporary dynamic systems models in an interesting way, noteworthy in terms of the neurorobotics research mentioned in the fourth section of this paper.

  8. He identifies a similar loop arising in the context of gameplay, as the significance of any given piece is partly determined by any single game piece and partly by the position of that piece relative other pieces, with all of this only significant in the context of possible future moves as constrained by the structured goal conditions of the game. For a heuristic chess program, for example, to escape such circularity, the interdependence of such definitions requires that some be fixed, at which point its limits “become a matter of principle rather than simply of practice” (1965, p. 73).

  9. For an interesting use of this example, see Williams et al. (1974).

  10. Problems accessible to “associationistic” processes include term-replacement (“mechanical dictionary”), simple memory (associating symbols with patterns), conditioned response (as a means to producing these associations) and trial and error (maze navigation) type problems, all of which are accessible to traditional symbol-pushing AI in the form of decision trees or list searches. The second type is also fully accessible to traditional AI. “Simple formal” type problems are fully governed by explicit rules defining contexts simple enough for meanings to be made completely explicit independent of context, including games whose moves can be explicitly counted out, theorem proofs involving searches for optimal applications of rules under explicit constraints, and pattern recognition given salient features are defined beforehand. The third type, “complex formal” problems, are only partially accessible to traditional AI, including “uncomputable” games whose decision trees cannot be brute-force counted-out including chess and go which require an appreciation of global context to settle on which features are essential and which not before deciding on further moves, and pattern recognition wherein regularities must first be discovered before similarities are recognized. Finally, Dreyfus holds that the fourth type of problem is wholly outside of machine comprehension on principle, consisting in “non-formal” problems with solutions requiring context sensitivity and an ability to distinguish essential from nonessential aspects in patterns which are distorted, the translation of natural language into metaphorical contexts, and the ability to structure otherwise seemingly unstructured problems through what Dreyfus calls “perceptive guess” (see the section headed “Areas of Intelligent Activity Classified with Respect to the Possibility of Artificial Intelligence in Each” beginning on p. 75).

  11. Alexander VonSchoenborn taught this lesson through the image of an ever-expanding sphere of light around an increasingly bright candle. The greater the light, the larger the unlit fringe.

  12. By “resistor analogue”, Dreyfus is characterizing a machine that works as do human beings to infer best explanations or to abduce optimal solutions such as we see in contemporary machine learning.

  13. Whereby AI becomes more intelligent than human beings, human beings are compelled to increasingly integrate themselves with AI to keep up with technological progress, and once bent to the solution of humanity’s most pressing problems, achieve a capacity to deliver humanity from them, cf. Kurzweil, 2005. For a technical argument as to why such an “explosion” of “superintelligence” in self-modifying AI is unlikely, see Benthall (2017).

  14. It is interesting to contrast these efforts with a related assessment of the “real danger of artificial intelligence”, that increasing automation by way of semi-autonomous machines promises to exacerbate rather than abate an historically unprecedented degree of wealth inequality (cf. Lee 2017) which has proven historically to precede mass violence, social upheaval, and in the twentieth century World War, and of which “autonomous weapons” will surely now play a part.

  15. Meanwhile, the simple-mindedness of such reasoning has not gone unrecognized by proponents of current research vectors. For example, late in 2015 the Information Technology and Innovation Foundation nominated the “loose coalition of scientists and luminaries who stirred fear and hysteria in 2015 by raising alarms that artificial intelligence (AI) could spell doom for humanity” for its annual Luddite Award, an honor that was finally awarded them in January of the next year (cf. Atkinson 2015; ITIF 2016).

  16. Noting that the cognizance of the individual of its situation within a social collective is “emergent” in 2001, Sun remarked on research in “bridging” the two (an image also central to da Costa and Pereira 2015) that “It helps to unearth the social unconscious embedded into social institutions and lodged inside individual minds” with the “crucial step in establishing the micro–macro link … between individual agents and society” becoming “how we should understand and characterize the structures of the social unconscious and its ‘‘emergence’’, computationally or otherwise.” (Sun 2001, page 2).

  17. This is the thesis driving the Bloom et al. (2017), for example.

References

  • Armstrong S, Bostrom N, Shulman C (2016) Racing to the precipice: a model of artificial intelligence development. AI Soc 31(2):201–206

    Article  Google Scholar 

  • Atkinson RD (2015) The 2015 ITIF Luddite Award Nominees: The Worst of the Year’s Worst Innovation Killers. http://www2.itif.org/2015-itif-luddite-award.pdf. Accessed 25 Nov 2017

  • Benthall S (2017) Do not fear the reaper: refuting bostrom’s superintelligence argument. Preprint, arXiv:1702.08495. https://arxiv.org/pdf/1702.08495.pdf. Accessed 25 Nov 2017

  • Bloom N, Jones CI, Reenen JV, Webb M (2017). Are ideas getting harder to find? Natl Bur Econ Res. http://www.nber.org/papers/w23782. Accessed 25 Nov 2017

  • Bostrom N (2012) The superintelligent will: motivation and instrumental rationality in advanced artificial agents. Mind Mach 22(2):71–85

    Article  Google Scholar 

  • Bostrom N (2013) Superintelligence: paths, dangers, strategies. Oxford University Press, Oxford

    Google Scholar 

  • Brey P (2001) Hubert Dreyfus—human versus machine. In: Achterhuis H, Crease RP (eds) American philosophy of technology: the empirical turn. Indiana University Press, Bloomington, pp 37–63

    Google Scholar 

  • Chalmers DJ (2010) The singularity: a philosophical analysis. J Conscious Stud 17:7–65

    Google Scholar 

  • da Cardoso CF, Pereira LM (2015) The emergence of artificial autonomy: a view from the foothills of a challenging. In: White J, Searle R (eds) Rethinking machine ethics in the age of ubiquitous technology. IGI Global, Hauppage

    Google Scholar 

  • Dreyfus HL (1995) Heidegger on gaining a free relation to technology. In: Feenburg A, Hannay A (eds) Technology and the politics of knowledge. Indiana University Press, Bloomington, pp 25–33

    Google Scholar 

  • Dreyfus HL (2007) Why heideggerian ai failed and how fixing it would require making it more heideggerian. Artif Intell 171(18):1137–1160

    Article  Google Scholar 

  • Dreyfus HL (2012) A history of first step fallacies. Mind Mach 22(2):87–99

    Article  Google Scholar 

  • Dreyfus HL, Rand Corp Santa Monica Calif (1965) Alchemy and artificial intelligence. Defense Technical Information Center, Ft. Belvoir. https://www.rand.org/content/dam/rand/pubs/papers/2006/P3244.pdf. Accessed 25 Nov 2017

  • Future of Life Institute (2015) AI open letter. https://futureoflife.org/ai-open-letter/. Accessed 25 Nov 2017

  • Future of Life Institute (2017) An open letter to the united nations convention on certain conventional weapons. https://futureoflife.org/autonomous-weapons-open-letter-2017. Accessed 26 Nov 2017

  • Harris S (2016) TEDSummit J 2016, can we build AI without losing control of it? TED Summit, June 2016. https://www.ted.com/talks/sam_harris_can_we_build_ai_without_losing_control_over_it/transcript. Accessed 25 Nov 2017

  • ITIF (2016) Artificial intelligence alarmists win ITIF’s annual luddite award. https://itif.org/publications/2016/01/19/artificial-intelligence-alarmists-win-itif’s-annual-luddite-award. Accessed 25 Nov 2017

  • Kenaw S (2008) Hubert L. Dreyfus’ critique of classical AI and its rationalist assumptions. Mind Mach 18(2):227–238

    Article  Google Scholar 

  • Kurzweil R (2005) The singularity is near: when humans transcend biology. Viking, New York, NY

    Google Scholar 

  • Lee K (2017) The real threat of artificial intelligence. https://www.nytimes.com/2017/06/24/opinion/sunday/artificial-intelligence-economic-inequality.html. Accessed 25 Nov 2017

  • Musk E (2014) 2014 MIT AeroAstro Centennial Symposium. http://aeroastro.mit.edu/aeroastro100/centennial-symposium. Accessed 25 Nov 2017

  • Musk E (2017) May be initiated not by the country leaders, but one of the AI’s, if it decides that a preemptive strike is most probable path to victory. https://twitter.com/elonmusk/status/904639405440323585. Accessed 25 Nov 2017

  • Pereira LM, Saptawijaya A (2015) Bridging two realms of machine ethics. In: White J, Searle R (eds) Rethinking machine ethics in the age of ubiquitous technology. IGI Global, Hauppage

    Google Scholar 

  • Pereira LM, Saptawijaya A (2016) Modeling collective morality via evolutionary game theory. In: Magnani L (ed) Programming machine ethics. Studies in applied philosophy, epistemology and rational ethics. Springer, Berlin, pp 26

    Google Scholar 

  • Putin V (2017) ‘Whoever leads in AI will rule the world’: Putin to Russian children on knowledge day. https://www.rt.com/news/401731-ai-rule-world-putin/. Accessed 25 Nov 2017

  • Rouse J (2013) What is conceptually articulated understanding? In: Schear JK (ed) Mind, reason, and being-in-the-world: the McDowell-Dreyfus debate. Routledge, London, pp 250–271

    Google Scholar 

  • Russell S, Dewey D, Tegmark M (2015) Research priorities for robust and beneficial artificial intelligence. AI Magaz 36(4):105–114

    Article  Google Scholar 

  • Silver D, Huang A, Maddison CJ, Guez A, Sifre L, Van DDG, Schrittwieser J et al (2016) Mastering the game of Go with deep neural networks and tree search. Nature 529(7587):484–489

    Article  Google Scholar 

  • Sun R (2001) Individual action and collective function: From sociology to multi-agent learning. Cogn Syst Res 2(1):1–3

    Article  Google Scholar 

  • Sun R (2013) Moral judgment, human motivation, and neural networks. Cognit Comput 5(4):566–579

    Article  MathSciNet  Google Scholar 

  • Sun R (2018) Intrinsic motivation for truly autonomous agents. In: Abbass H, Scholz J, Reid D (eds) Foundations of trusted autonomy. Springer, Berlin

    Google Scholar 

  • Sun R, Helie S (2015) Accounting for creativity within a psychologically realistic cognitive architecture. In: Besold T, Schorlemmer M, Smaill A (eds) Computational creativity research: towards creative machines. Atlantis thinking machines, 7th edn. Atlantis Press, Paris

    Google Scholar 

  • Sun R, Wilson N, Lynch M (2016) Emotion: a unified mechanistic interpretation from a cognitive architecture. Cognit Comput 8(1):1–14

    Article  Google Scholar 

  • Tani J (2016) Exploring robotic minds: actions, symbols, and consciousness as self-organizing dynamic phenomena. Oxford University Press, Oxford

    Book  Google Scholar 

  • White J (2016) Simulation, self-extinction, and philosophy in the service of human civilization. AI Soc 31(2):171–190

    Article  Google Scholar 

  • White J, Tani J (2017) From biological to synthetic neurorobotics approaches to understanding the structure essential to consciousness (part 3). APA Newslett Philos Comput 17(1):11–22

    Google Scholar 

  • Williams B, Haugeland J, Dreyfus H (1974) An exchange on artificial intelligence. http://www.nybooks.com/articles/1974/06/27/an-exchange-on-artificial-intelligence. Accessed 25 Nov 2017

Download references

Acknowledgements

The author wishes to thank Karamjit Gill for this journal and for his support, as well as for arranging anonymous review resulting in insightful directions for significant improvements. This work is dedicated to Hubert Dreyfus.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jeffrey White.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

White, J. Dreyfus on the “Fringe”: information processing, intelligent activity, and the future of thinking machines. AI & Soc 34, 301–312 (2019). https://doi.org/10.1007/s00146-018-0837-5

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00146-018-0837-5

Keywords

Navigation