Abstract
Erik Hollnagel’s body of work in the past three decades has molded much of the current research approach to system safety, particularly notions of “error”. Hollnagel regards “error” as a dead-end and avoids using the term. This position is consistent with Rasmussen’s claim that there is no scientifically stable category of human performance that can be described as “error”. While this systems view is undoubtedly correct, “error” persists. Organizations, especially formal business, political, and regulatory structures, use “error” as if it were a stable category of human performance. They apply the term to performances associated with undesired outcomes, tabulate occurrences of “error”, and justify control and sanctions through “error”. Although a compelling argument can be made for Hollnagel’s view, it is clear that notions of “error” are socially and organizationally productive. The persistence of “error” in management and regulatory circles reflects its value as a means for social control.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Abrams J, Branch J (2010) Fast and risky, sledding track drew red flags. NY Times 1:4
Adler-Milstein JA, Singer SJ, Toffel MW (2009) Operational failures and problem solving: an empirical study of incident reporting. Working paper 10–017. Harvard Business School, Canbridge
Bowker GC, Star SL (2000) Sorting things out: classification and its consequences. MIT Press, Cambridge
Cook RI, Woods DD (2007) Distancing through differencing: an obstacle to organizational learning following accidents. In: Hollnagel E, Woods DD, Leveson N (eds) Resilience engineering: concepts and precepts. Ashgate, Aldershot, pp 329–338
Frei FX, Kalakota R, Leone AJ, Marx LM (1999) Process variation as a determinant of banking performance: evidence from the retail banking industry. Mgt Sci 45(9):1210–1220
Hollnagel E (1983) Human error. Position paper for NATO conference on human error. Bellagio, Italy
Hollnagel E (1988) Mental models and model mentality. In: Goodstein LP, Andersen HB, Olsen SE (eds) Tasks, errors and mental models. Taylor and Francis, New York, pp 261–268
Hollnagel E (1993) Human reliability analysis: context and control. Academic Press, London
Hollnagel E (1998) Context, cognition and control. In: Waern Y (ed) Co-operative process management: cognition and information technology. Taylor and Francis, Bristol, pp 27–52
Hollnagel E (2004) Barrier analysis and accident prevention. Ashgate Publishing, Aldershot
Hollnagel E (2009) The ETTO principle: efficiency-thoroughness trade-off. Ashgate Publishing, Farnham
Hollnagel E, Woods D (1983) Cognitive systems engineering: new wine in new bottles. Int J Man Mach Stud 18:583–600
Hollnagel E, Woods DD (2005) Joint cognitive systems: foundations of cognitive systems engineering. Taylor and Francis/CRC Press, Boca Raton
Keeler LJ (1934) The problem of error from Plato to Kant: a historical and critical study. (Analecta Gregoriana, vol VI). Rome: Pontificia Universtas Gregoriana, p 174
Khatri N, Baveja A, Boren SA, Mammo A (2006) Medical errors and quality of care: from control to commitment. Harv Bus Rev
Margolis JD, Stoltz PG (2010) How to bounce back from adversity. Harv Bus Rev
Nemeth C (2005) Healthcare forensics. In: Karwowski W, Noy I (eds) Handbook of human factors forensics. Taylor and Francis, New York, pp 37:1–37:18
Nemeth C, Cook RI, Patterson E, Donchin Y, Rogers M, Ebright P (2004) Afterwords: the quality of medical accident investigations and analyses. In: Human factors and ergonomics society annual meeting, New Orleans
Nemeth CP, Nunnally ME, O’Connor MF, Klock PA, Cook RI (2005) Making information technology a team player in safety: the case of infusion devices. In: Henricksen K, Battles J, Marks E, Lewin D (eds) Advances in patient safety: from research to implementation, vol 1. Agency for Healthcare Research and Quality, Washington, DC, pp 319–330
Nemeth C, Dierks M, Patterson E Donchin Y, Crowley J, McNee S, Powell T, Cook RI (2006) Learning from Investigation. In: Proceedings of the human factors and ergonomics society annual meeting, San Francisco, pp 914–917
Nunnally M, Nemeth C, Brunetti V, Cook R (2004) Lost in menuspace: user interactions with complex medical devices. In: Nemeth C, Cook R, Woods (eds) Special issue on studies in healthcare technical work. IEEE transactions on systems, man and cybernetics part A, vol 34, no 6, pp 736–742
Patterson ES, Cook RI, Render ML, Woods DD (2001) New Arctic air crash aftermath role-play simulation: orchestrating fundamental surprise. In: Proceedings of the human factors and ergonomics society 45th annual meeting, Minneapolis, MN
Rasmussen J (1986) Information processing and human-machine interaction: an approach to cognitive engineering. New York, North-Holland
Rasmussen J (1990) The role of error in organizing behavior. Ergonomics 33:1185–1199
Rasmussen J (2000) The concept of error: is it useful for the design of safe systems in Healthcare? In: Vincent C, de Mol B (eds) Safety in medicine. Pergamon, London, pp 31–48
Reason J (1997) Managing the risk of organizational accidents. Ashgate Publishing, Aldershot
Simon HA (1996) The sciences of the artificial. MIT Press, Cambridge
Stoltz PG (2000) The adversity quotient @ work. William Morrow and Company, New York
Tasca L (1990) Unpublished doctoral dissertation. The social construction of human error, State University of New York at Stonybrook
Weick KE (2001) Tool retention and fatalities in wildland fire settings: conceptualizing the naturalistic. In: Salas E, Klein G (eds) Linking expertise and naturalistic decision making. Lawrence Erlbaum Associates, Mahwah, pp 321–336
Woods DD (1988) Coping with complexity: the psychology of human behavior in complex systems. In: Goodstein LP, Andersen HB, Olsen SE (eds) Tasks, errors, and mental models. Taylor & Francis, New York
Woods DD, Cook RI (2002) Nine steps to move forward from error. Cogn Tech Work 4:137–144
Woods DD, Hollnagel E (1987) Mapping cognitive demands in complex problem-solving worlds. Int J Man Mach Stud 26:257–275
Acknowledgments
Research conducted during the MEDCAS project was funded by support from the VHA Foundation and the US Food and Drug Administration Center for Devices and Radiologic Health. Elements of this paper are based on Cook, RI (2002, September) The Uses of Error: A Reply to Senders, an unpublished position paper for the Clambake III Conference, University of Chicago Gleacher Center, Chicago, IL.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Cook, R.I., Nemeth, C.P. “Those found responsible have been sacked”: some observations on the usefulness of error. Cogn Tech Work 12, 87–93 (2010). https://doi.org/10.1007/s10111-010-0149-0
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10111-010-0149-0