Abstract
Previous researches show that people are only able to recognise deception with an accuracy of 54%. In this paper, using physiological signals from observers, we train a three-layer neural network, a long short-term memory (LSTM) and a multi-tasking learning neural network (MTL-NN). We demonstrate that examined models are able to identify deception with an accuracy up to 62%, surpassing the average accuracy of human deception detection. The superior deception recognition ability shows that these tools are capable of helping people discriminate against deception. Further, to improve the interpret-ability of neural networks, we extract rules from the trained models using sensitivity analysis. We find that the rule extraction methods using sensitivity analysis along with genetic algorithm (GA) based data reduction successfully explain all three neural network models. We hope the rule extraction methods can help to improve the interpret-ability of neural networks.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Andrews, R., Diederich, J., Tickle, A.B.: Survey and critique of techniques for extracting rules from trained artificial neural networks. Knowl.-Based Syst. 8(6), 373–389 (1995)
Chen, Z., Badrinarayanan, V., Lee, C.Y., Rabinovich, A.: Gradnorm: gradient normalization for adaptive loss balancing in deep multitask networks. In: International Conference on Machine Learning, pp. 794–803. PMLR (2018)
Cho, K., et al.: Learning phrase representations using rnn encoder-decoder for statistical machine translation. arXiv preprint arXiv:1406.1078 (2014)
Choi, J.Y., Choi, C.H.: Sensitivity analysis of multilayer perceptron with differentiable activation functions. IEEE Trans. Neural Netw. 3(1), 101–107 (1992)
Cleveland, W.S.: Robust locally weighted regression and smoothing scatterplots. J. Am. Stat. Assoc. 74(368), 829–836 (1979)
Elkins, A., Zafeiriou, S., Pantic, M., Burgoon, J.: Unobtrusive deception detection. In: The Oxford Handbook of Affective Computing, pp. 503–515. Oxford University Press (2014)
Gedeon, T., Turner, H.S.: Explaining student grades predicted by a neural network. In: Proceedings of 1993 International Conference on Neural Networks (IJCNN-93-Nagoya, Japan), vol. 1, pp. 609–612 (1993). https://doi.org/10.1109/IJCNN.1993.713989
Guo, M., Haque, A., Huang, D.-A., Yeung, S., Fei-Fei, L.: Dynamic task prioritization for multitask learning. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11220, pp. 282–299. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01270-0_17
Hailesilassie, T.: Rule extraction algorithm for deep neural networks: a review. arXiv preprint arXiv:1610.05267 (2016)
Hashem, S.: Sensitivity analysis for feedforward artificial neural networks with differentiable activation functions. In: [Proceedings 1992] IJCNN International Joint Conference on Neural Networks, vol. 1, pp. 419–424. IEEE (1992)
Mitchell, M.: An Introduction to Genetic Algorithms. MIT press, Cambridge (1998)
Rodner, E., Simon, M., Fisher, R.B., Denzler, J.: Fine-grained recognition in the noisy wild: sensitivity analysis of convolutional neural networks approaches. arXiv preprint arXiv:1610.06756 (2016)
Safavian, S.R., Landgrebe, D.: A survey of decision tree classifier methodology. IEEE Trans. Syst. Man Cybern. 21(3), 660–674 (1991)
Zhu, X., Gedeon, T., Caldwell, S., Jones, R., Gu, X.: Deceit detection: identification of presenter’s subjective doubt using affective observation neural network analysis. In: 2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC), pp. 3174–3181. IEEE (2020)
Zhu, X., Qin, Z., Gedeon, T., Jones, R., Hossain, M.Z., Caldwell, S.: Detecting the Doubt Effect and Subjective Beliefs using neural networks and observers’ pupillary responses. In: Cheng, L., Leung, A.C.S., Ozawa, S. (eds.) ICONIP 2018. LNCS, vol. 11304, pp. 610–621. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-04212-7_54
Zilke, J.R., Loza Mencía, E., Janssen, F.: DeepRED – rule extraction from deep neural networks. In: Calders, T., Ceci, M., Malerba, D. (eds.) DS 2016. LNCS (LNAI), vol. 9956, pp. 457–473. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46307-0_29
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Zhang, X., Zhu, X. (2021). Explaining Neural Network Results by Sensitivity Analysis for Deception Detection. In: Mantoro, T., Lee, M., Ayu, M.A., Wong, K.W., Hidayanto, A.N. (eds) Neural Information Processing. ICONIP 2021. Communications in Computer and Information Science, vol 1517. Springer, Cham. https://doi.org/10.1007/978-3-030-92310-5_15
Download citation
DOI: https://doi.org/10.1007/978-3-030-92310-5_15
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-92309-9
Online ISBN: 978-3-030-92310-5
eBook Packages: Computer ScienceComputer Science (R0)