Abstract
Engineering deals with different problem situations than science, and theories in engineering are different to theories in science. So, the growth of knowledge in engineering is also different to that in science. Nonetheless, methodological issues in engineering epistemology can be explored by adapting frameworks already established in the philosophy of science. In this paper I use critical rationalism and Popper’s three worlds framework to investigate error elimination and the growth of knowledge in engineering. I discuss engineering failure arising from the falsification of engineering theories, and present taxonomies of the sources of falsification and responses to falsification in engineering. From this I discuss contexts of research and design in engineering, ad hoc rescue of engineering theories, and engineering assurance.
Similar content being viewed by others
Notes
Brooks (1996, p. 62) says of this: “A high-energy physicist may easily spend most of his time building his apparatus; a spacecraft engineer may easily spend most of his time studying the behavior of materials in vacuum. Nevertheless, the scientist builds in order to study; the engineer studies in order to build.”
The appropriate response in this situation is not always initially clear, but ultimately one must either reject the general theory, or reject the formalisation of the new observation.
Of course, a specific theory may turn out to be false, in which case it is rejected and a new tentative theory may be proposed to replace it. See Sect. 4.1 for a discussion of ad hoc rescue of theories.
Not every artefact failure will falsify an engineering theory. Often, acceptable requirements for use are qualified by probabilistic reliability conditions. This allows the use of imprecise engineering theories that accommodate occasional failure, arising for example from inevitable variations in quality of materials used in the construction of artefacts.
Failures can be false negatives (where a bad artefact is mistakenly not identified as such) or false positives (where a good artefact is mistakenly thought to be bad). A false negative during design will not necessarily lead to a failure in use, because the artefact may satisfy its requirements in all of the specific situations in which it actually used, even though it would not function correctly in all of the specific situations it was required to be able to be used. False positives do not typically lead directly to engineering failures because the artefacts are in reality good, and as the artefacts are incorrectly deemed unsuitable, they are typically not used anyway. As well as false negatives and false positives, engineering theories may simply fail to show whether an artefact will meet its requirements. That is, theories may be indeterminate in some situations. However, like false-positive situations, indeterminate analyses do not lead to artefacts being deemed suitable for use, and so typically do not directly lead to engineering failures. In this paper I mostly focus on false-negative situations.
Some readers may wonder why an example from computing is relevant to a paper on engineering. The short answer is that computer systems engineering (including software engineering) is, from a methodological perspective, part of engineering. As argued in the previous paper (Staples 2014), computer programs as written embody a kind of objective knowledge, in World 3. Nonetheless, computer programs, when executing, are physical processes in World 1. They execute on physical hardware, and their execution leads to physical changes in usage situations. Indeed the example of TEMPEST attacks strongly supports the position that computer systems (combining hardware and software) are engineered systems, because the attacks exploit physical characteristics of the computer systems. Software as a formal entity in computer science is the wrong category of thing to be subject to physical side-channel attacks. Engineering theories about the security of software-based systems must include operational conditions and constraints on hardware in order to avoid falsification by such attacks.
References
Anderson, R. (2008). Security engineering: A guide to building dependable distributed systems (2nd ed.). Indianapolis: Wiley.
Barber, E. H. E., Greenwood, J. N., & Matheson, J. A. L. (1963). Report of the royal commission into the failure of kings bridge. Technical report, Victorian Royal Commission into the Failure of Kings Bridge.
Bell, D. E., LaPadula, L. J. (1973). Secure computer systems: Mathematical foundations. Technical Report MTR-2547, MITRE.
Boon, M., & Knuuttila, T. (2009). Models as epistemic tools in engineering sciences. In A. Meijers (Ed.), Philosophy of technology and engineering sciences (Vol. 9, pp. 693–726)., Handbook of the philosophy of science Amsterdam: Elsevier.
Brooks, F. P, Jr. (1996). The computer scientist as toolsmith II. Communications of the ACM, 39(3), 61–68.
Cartwright, N. (1983). How the laws of physics lie. Oxford: Oxford University Press.
Clausen, J., & Cantwell, J. (2007). Reasoning with safety factor rules. Techné: Research in Philosophy and Technology, 11(1), 55–70.
Constant, II, E. W. (1984). Communities and hierarchies: Structure in the practice of science and technology. In Laudan (1984b), pp. 27–46.
Constant, E. W, I. I. (1999). Reliable knowledge and unreliable stuff. Technology and Culture, 40(2), 324–357.
Davis, M. (2010). Distinguishing architects from engineers: A pilot study in differences between engineers and other technologists. In I .van de Poel & D. E.Goldberg (Eds.), Philosophy and engineering: An emerging agenda (pp. 15–30), Vol. 2 of Philosophy of engineering and technology. Dordrecht: Springer.
Ferguson, E. S. (1992). Engineering and the mind’s eye. Cambridge, MA: The MIT Press.
Gainsburg, J., Rodriquez-Lluesma, C., & Bailey, D. E. (2010). A “knowledge profile” of an engineering occupation: Temporal patterns in the use of engineering knowledge. Engineering Studies, 2(3), 197–219.
Helton, J. C., & Oberkampf, W. L. (2004). Alternative representations of epistemic uncertainty. Reliability Engineering and System Safety, 85, 1–10.
Hoare, C. A. R. (1996). The logic of engineering design. Microprocessing and Microprogramming, 41, 525–539.
Houkes, W., & Vermaas, P. E. (2009). Produced to use: Combining two key intuitions on the nature of artefacts. Techné: Research in Philosophy and Technology, 13(2), 123–136.
Hughes, T. P. (1976). The science-technology interaction: The case of high-voltage power transmission systems. Technology and Culture, 17(4), 646–662.
Johnson, A. (2009). Hitting the brakes: Engineering design and the production of knowledge. Durham, NC: Duke University Press.
Kroes, P. (2002). Design methodology and the nature of technical artefacts. Design Studies, 23, 287–302.
Lakatos, I. (1970). Falsification and the methodology of scientific research programmes. In I. Lakatos & A. Musgrave (Eds.), Criticism and the growth of knowledge (pp. 91–196). London: Cambridge University Press.
Lakatos, I. (1976). Proofs and refutations. Cambridge: Cambridge University Press.
Laudan, R. (1984). Cognitive change in technology and science. In R. Laudan (Ed.), The Nature of Technological Knowledge. Dordrecht: D. Reidel.
Laudan, R. (Ed.). (1984b). The Nature of Technological Knowledge. Dordrecht: D. Reidel.
Laymon, R. (1989). Applying idealized scientific theories to engineering. Synthese, 81, 353–371.
Layton, E. (1971). Mirror-image twins: The communities of science and technology in 19th-century America. Technology and Culture, 12(4), 562–580.
MacKenzie, D. (2001). Mechanizing proof: Computing, risk, and trust. Cambridge, MA: The MIT Press.
Marshall, R. D., Pfrang, E. O., Leyendecker, E. V., Woodward, K. A., Reed, R. P., Kasen, M. B., Shives, T. R. (1982). Investigation of the kansas city hyatt regency walkways collapse. Technical Report 143, U.S. Dept. of Commerce, National Bureau of Standards.
Mayo, D. G. (1996). Error and the growth of experimental knowledge. Chicago: The University of Chicago Press.
McLean, J. (1985). A comment on the “basic security theorem” of Bell and LaPadula. Information Processing Letters, 20, 67–70.
Oddie, G. (1981). Verisimilitude reviewed. The British Journal for the Philosophy of Science, 32, 237–265.
Petroski, H. (1996). Invention by design: How Engineers get from thought to thing. Cambridge, MA: Harvard University Press.
Petroski, H. (2012). To forgive design: Understanding failure. Cambridge, MA: The Belknap Press of Harvard University Press.
Pirtle, Z. (2010). How the models of engineering tell the truth. In I. van de Poel & D. E. Goldberg (Eds.), Philosophy and engineering: An emerging agenda (pp. 95–108), Vol. 2 of Philosophy of engineering and technology. Dordrecht: Springer.
Polanyi, M. (1958). Personal knowledge: Towards a post-critical philosophy. London: Routledge.
Pols, A. (2010). Transferring responsibility through use plans. In I. van de Poel & D. E. Goldberg (Eds.), Philosophy and engineering: An emerging agenda (pp. 189–203), Vol. 2 of Philosophy of engineering and technology. Dordrecht: Springer.
Popper, K. R. (1959). The logic of scientific discovery (3rd ed.). London: Routledge. printed 2002.
Popper, K. R. (1963). Conjectures and Refutations (2nd ed.). London: Routledge. printed 2002.
Popper, K. R. (1972). Objective knowledge: An evolutionary approach. New York: Oxford University Press.
Popper, K. R. (1977). The worlds 1, 2 and 3. In K. R. Popper & J. C. Eccles (Eds.), The Self and its brain: An argument for interactionism (pp. 36–50). London: Routledge.
Popper, K. R. (1978). Three worlds. The Tanner Lecture on Human Values. Accessed January 18, 2014, from http://tannerlectures.utah.edu/_documents/a-to-z/p/popper80.pdf.
Rapp, F. (1981). Analytical philosophy of technology. Dordrecht: D. Reidel.
Rittel, H. (1972). On the planning crisis: Systems analysis of the ‘first and second generations’. Bedriftsøkonomen, 8, 390–396.
Rushby, J. (2013). Mechanized support for assurance case argumentation. In Proceedings of the 1st International workshop on argument for agreement and assurance. London: Springer.
Staples, M. (2014). Critical rationalism and engineering: ontology. Synthese, 191(10), 2255–2279.
van de Poel, I. (2010). Philosophy and engineering: Setting the stage. In I. van de Poel & D. E. Goldberg (Eds.), Philosophy and engineering: An emerging agenda (pp. 1–11), Vol. 2 of Philosophy of engineering and technology. Dordrecht: Springer.
Vermaas, P. E. (2010). Focussing philosophy of engineering: Analyses of technical functions and beyond. In van de Poel and Goldberg (Eds.), Philosophy and engineering: An emerging agenda (pp. 61–72), Vol. 2 of Philosophy of engineering and technology. Dordrecht: Springer.
Vincenti, W. (1990). What engineers know and how they know it. Baltimore, MD: John Hopkins University Press.
Wimsatt, W. C. (2007). False models as means to truer theories. In Re-engineering philosophy for limited beings: Piecewise approximations to reality. Cambridge, MA: Harvard University Press.
Acknowledgments
NICTA is funded by the Australian Government through the Department of Communications and the Australian Research Council through the ICT Centre of Excellence Program.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Staples, M. Critical rationalism and engineering: methodology. Synthese 192, 337–362 (2015). https://doi.org/10.1007/s11229-014-0571-6
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11229-014-0571-6