ABSTRACT
Deep learning in computer vision has shown remarkable success in the performance of detection systems for plant diseases. However, due to the complexity and deeply nested structure of these models, these are still considered as black-box and explanations are not intuitive for human users. Many researchers have developed deep neural architectures for plant disease detection but have not provided classification explanations. To be used in practical applications, our model needs to explain why the model classified a given image. Explainable Artificial Intelligence (XAI) provides algorithms that can generate human-understandable explanations of AI decisions. In this paper, we summarize recent developments in XAI techniques, develop a plant disease detection system, and most importantly an explainable AI method named Gradient-weighted Class Activation Mapping ++ (GradCAM++) is used to locate the disease and highlight the most important regions on the leaves contributing towards the classification.
- Amina Adadi and Mohammed Berrada. 2018. Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI). IEEE Access 6(2018), 52138–52160. 10.1109/access.2018.2870052; https://dx.doi.org/10.1109/access.2018.2870052Google ScholarCross Ref
- Ajaya Adhikari, David Tax, Riccardo Satta, and Matthias Faeth. 2019. LEAFAGE: Example-based and Feature importance-based Explanations for Black-box ML models. 1–7. 10.1109/FUZZ-IEEE.2019.8858846Google Scholar
- Rishabh Agarwal, Nicholas Frosst, Xuezhou Zhang, Rich Caruana, and Geoffrey E. Hinton. 2020. Neural Additive Models: Interpretable Machine Learning with Neural Nets. CoRR abs/2004.13912(2020).Google Scholar
- Marko Arsenovic, Mirjana Karanovic, Srdjan Sladojevic, Andras Anderla, and Darko Stefanovic. 2019. Solving Current Limitations of Deep Learning Based Approaches for Plant Disease Detection. Symmetry 11, 7 (2019), 939–939. 10.3390/sym11070939; https://dx.doi.org/10.3390/sym11070939Google ScholarCross Ref
- Sebastian Bach, Alexander Binder, Grégoire Montavon, Frederick Klauschen, Klaus-Robert Müller, and Wojciech Samek. 2015. On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation. PLOS ONE 10, 7 (2015), e0130140–e0130140. 10.1371/journal.pone.0130140; https://dx.doi.org/10.1371/journal.pone.0130140Google ScholarCross Ref
- O Biran and C Cotton. 2017. Explanation and justification in machine learning: A survey. IJCAI-17 workshop on explainable AI (XAI) 8 (2017), 1–47.Google Scholar
- D Blancard. 2012. 2 - Diagnosis of Parasitic and Nonparasitic Diseases. Academic Press, The Netherlands.Google Scholar
- Davide Castelvecchi. 2016. Can we open the black box of AI?Nature News 538, 7623 (2016), 20–20.Google Scholar
- A Chattopadhay, A Sarkar, P Howlader, and V N Balasubramanian. 2018. Grad-CAM++: Generalized Gradient-Based Visual Explanations for Deep Convolutional Networks. Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV)(2018), 12–15.Google ScholarCross Ref
- J Chen, Q Liu, and L Gao. 2019. Visual Tea Leaf Disease Recognition Using a Convolutional Neural Network Model.10.3390/sym11030343; https://www.mdpi.com/2073-8994/11/3/343Google Scholar
- David Danks and Alex John London. 2017. Regulating Autonomous Systems: Beyond Standards. IEEE Intelligent Systems 32, 1 (2017), 88–91. 10.1109/mis.2017.1; https://dx.doi.org/10.1109/mis.2017.1Google ScholarDigital Library
- David Danks and Alex John London. 2017. Regulating Autonomous Systems: Beyond Standards. IEEE Intelligent Systems 32, 1 (2017), 88–91. 10.1109/mis.2017.1; https://dx.doi.org/10.1109/mis.2017.1Google ScholarDigital Library
- [13] F Doshi-Velez and B Kim.2017.Google Scholar
- H Durmuş, E O Güneş, and M Kırcı. 2017. Disease detection on the leaves of the tomato plants by using deep learning. 2017 6th International conference on Agro-Geoinformatics (2017), 1–5.Google ScholarCross Ref
- Konstantinos P. Ferentinos. 2018. Deep learning models for plant disease detection and diagnosis., 311-318 pages. 10.1016/j.compag.2018.01.009; https://dx.doi.org/10.1016/j.compag.2018.01.009Google ScholarDigital Library
- R Fergus. 2012. Deep learning methods for vision.Google Scholar
- Geetharamani G. and Arun Pandian J.2019. Identification of plant leaf diseases using a nine-layer deep convolutional neural network. Computers & Electrical Engineering 76 (2019), 323–338. 10.1016/j.compeleceng.2019.04.011; https://dx.doi.org/10.1016/j.compeleceng.2019.04.011Google ScholarDigital Library
- A Ghorbani, J Wexler, J Y Zou, and B Kim. 2019. Towards automatic concept-based explanations. Advances in Neural Information Processing Systems (2019), 9273–9282.Google Scholar
- L H Gilpin, D Bau, B Z Yuan, A Bajwa, M Specter, and L. 2018. Explaining Explanations: An Overview of Interpretability of. Machine Learning (2018).Google Scholar
- Riccardo Guidotti, Anna Monreale, Salvatore Ruggieri, Dino Pedreschi, Franco Turini, and Fosca Giannotti. 2018. Local Rule-Based Explanations of Black Box Decision Systems. CoRR abs/1805.10820(2018).Google Scholar
- Riccardo Guidotti, Anna Monreale, Salvatore Ruggieri, Franco Turini, Fosca Giannotti, and Dino Pedreschi. 2019. A Survey of Methods for Explaining Black Box Models. ACM Computing Surveys 51, 5 (2019), 1–42. 10.1145/3236009; https://dx.doi.org/10.1145/3236009Google ScholarDigital Library
- D. Gunning. [n.d.]. Explainable artificial intelligence (XAI)., 2017-2017 pages.Google Scholar
- David Gunning and David Aha. 2019. DARPA’s Explainable Artificial Intelligence (XAI) Program. AI Magazine 40 (06 2019), 44–58. 10.1609/aimag.v40i2.2850Google Scholar
- I M Hanssen and M Lapidot. 2012. Major tomato viruses in theMediterranean basin. Adv. Virus Res 84(2012), 31–66.Google ScholarCross Ref
- Andreas Holzinger, Georg Langs, Helmut Denk, Kurt Zatloukal, and Heimo Müller. 2019. Causability and explainability of artificial intelligence in medicine. WIREs Data Mining and Knowledge Discovery 9, 4 (2019), 1312–1312. 10.1002/widm.1312; https://dx.doi.org/10.1002/widm.1312Google ScholarCross Ref
- D Hughes and M Salathe. 2015. An open access repository of images on plant health to enable the development of mobile disease diagnostics., 13 pages.Google Scholar
- M M Islam, M R Haque, and H Iqbal. 2020. Breast Cancer Prediction: A Comparative Study Using Machine Learning Techniques. SN COMPUT. SCI 1(2020), 290–290.Google ScholarDigital Library
- Peng Jiang, Yuehan Chen, Bin Liu, Dongjian He, and Chunquan Liang. 2019. Real-Time Detection of Apple Leaf Diseases Using Deep Learning Approach Based on Improved Convolutional Neural Networks. IEEE Access 7(2019), 59069–59080. 10.1109/access.2019.2914929; https://dx.doi.org/10.1109/access.2019.2914929Google ScholarCross Ref
- K Kamal, Z Yin, M Wu, and Z Wu. 2019. Depthwise separable convolution architectures for plant disease classification. Comput. Electron. Agric 165 (2019).Google Scholar
- Asifullah Khan, Anabia Sohail, Umme Zahoora, and Aqsa Saeed Qureshi. 2020. A survey of the recent architectures of deep convolutional neural networks. Artificial Intelligence Review 53, 8 (2020), 5455–5516. 10.1007/s10462-020-09825-6; https://dx.doi.org/10.1007/s10462-020-09825-6Google ScholarDigital Library
- J A Kroll, J Huey, S Barocas, J R Edwardw, D G Felten, H Reiden-Berg, and Yu Robinson. 2017. Accountable algorithms.U. Penn. Law Rev 165(2017), 633–705.Google Scholar
- Zachary C. Lipton. 2018. The Mythos of Model Interpretability., 31-57 pages. 10.1145/3236386.3241340; https://dx.doi.org/10.1145/3236386.3241340Google ScholarDigital Library
- W Liu, Z Wang, and X Liu. 2017. A survey of deep neural network architectures and their applications. Neurocomputing 234(2017), 11–26.Google ScholarCross Ref
- S M Lundberg and S I Lee. 2017. A unified approach to interpreting model predictions. Proceedings of the Advances in Neural Information Processing Systems (2017), 4765–4774.Google Scholar
- S V Militante, B D Gerardo, and N V Dionisio. 2019. Plant Leaf Detection and Disease Recognition using Deep Learning. 2019 IEEE Eurasia Conference on IOT, Communication and Engineering (2019), 579–582.Google ScholarCross Ref
- T Miller. 2019. Explanation in artificial intelligence: Insights from the social sciences. Artif. Intell 267(2019), 1–38.Google ScholarCross Ref
- J Ni, Y Chen, Y Chen, J Zhu, D Ali, and W Cao. 2020. A Survey on Theories and Applications for Self-Driving Cars Based on. Deep Learning Methods. Appl. Sci 10 (2020), 2749–2749.Google Scholar
- V Ponnaganti, M Moh, and T Moh. 2020. Deep Learning for LiDAR-Based Autonomous Vehicles in Smart Cities. In Handbook of Smart Cities, Augusto J.C. (Ed.). Springer.Google Scholar
- Karthik R., Hariharan M., Sundar Anand, Priyanka Mathikshara, Annie Johnson, and Menaka R.2020. Attention embedded residual CNN for disease detection in tomato leaves. Applied Soft Computing 86 (2020), 105933–105933. 10.1016/j.asoc.2019.105933; https://dx.doi.org/10.1016/j.asoc.2019.105933Google ScholarDigital Library
- A Ramcharan, P Mccloskey, K Baranowski, N Mbilinyi, L Mrisho, M Ndalahwa, J Legg, and D Hughes. 2019. A mobile-based deep learning model for cassava disease diagnosis. Front Plant Sci 10(2019), 272–272.Google ScholarCross Ref
- S S Raoof, M A Jabbar, and S A Fathima. 2020. Lung Cancer Prediction using Machine Learning: A Comprehensive Approach. In 2020 2nd International Conference on Innovative Mechanisms for Industry Applications (ICIMIA). 108–115.Google ScholarCross Ref
- T Ribeiro, S Singh, and C Guestrin. 2016. Why should I trust you?” Explaining the predictions of any classifier. Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining(2016), 1135–1144.Google ScholarDigital Library
- T Ribeiro, S Singh, and C Guestrin. 2018. Anchors: High-precision model-agnostic explanations. Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence (2018).Google ScholarCross Ref
- Saleem, Potgieter, and Mahmood Arif. 2019. Plant Disease Detection and Classification by Deep Learning., 468-468 pages. 10.3390/plants8110468; https://dx.doi.org/10.3390/plants8110468Google Scholar
- R R Selvaraju, M Cogswell, A Das, R Vedantam, D Parikh, and D Batra. 2017. Grad-cam: Visual explanations from deep networks via gradient-based localization. Proceedings of the IEEE International Conference on Computer Vision (2017), 618–626.Google ScholarCross Ref
- S T Shane, T Mueller, R R Hoffman, W Clancey, and G Klein. 2019. Explanation in Human-AI Systems: A Literature Meta-Review Synopsis of Key Ideas and Publications and Bibliography for Explainable AI. Defense Advanced Research Projects Agency (DARPA) XAI Program (2019).Google Scholar
- A Shrikumar, P Greenside, and A Kundaje. 2017. Learning Important Features Through Propagating Activation Differences. Proceedings of the 34th International Conference on Machine Learning 70(2017), 3145–3153.Google ScholarDigital Library
- F K Dosilovi´c, M Brci´c, and N Hlupi´c. 2018. Explainable artificial intelligence: A survey. 41st International Convention on Information and Communication Technology, Electronics and Microelectronics (MIPRO)(2018), 210–215.Google Scholar
- K Simonyan, A Vedaldi, and A Zisserman. 2014. Deep inside convolutional networks: Visualising image classification models and saliency maps.Google Scholar
- Karen Simonyan and Andrew Zisserman. 2015. Very Deep Convolutional Networks for Large-Scale Image Recognition.Google Scholar
- N. Sneha and Tarun Gangil. 2019. Analysis of diabetes mellitus for early prediction using optimal features selection. Journal of Big Data 6, 1 (2019), 13–13. 10.1186/s40537-019-0175-6; https://dx.doi.org/10.1186/s40537-019-0175-6Google ScholarCross Ref
- J T Springenberg, A Dosovitskiy, T Brox, and M Riedmiller. 2015. Striving for Simplicity: The All Convolutional Net.Google Scholar
- A Sundararajan, Q Taly, and Yan. 2017. Axiomatic attribution for deep networks. Proceedings of the 34th International Conference on Machine Learning 70(2017), 3319–3328.Google ScholarDigital Library
- E Tjoa and C Guan. 2019. A survey on explainable artificial intelligence (XAI): Towards medical XAI.Google Scholar
- Edna Chebet Too, Li Yujian, Sam Njuki, and Liu Yingchun. 2019. A comparative study of fine-tuning deep learning models for plant disease identification. Computers and Electronics in Agriculture 161 (2019), 272–279. 10.1016/j.compag.2018.03.032; https://dx.doi.org/10.1016/j.compag.2018.03.032Google ScholarDigital Library
- M. P. Vaishnnave, K. Suganya Devi, and P. Ganeshkumar. 2020. Automatic method for classification of groundnut diseases using deep convolutional neural network. Soft Computing 24, 21 (2020), 16347–16360. 10.1007/s00500-020-04946-0; https://dx.doi.org/10.1007/s00500-020-04946-0Google ScholarDigital Library
- Muhammad Rehman Zafar and Naimul Mefraz Khan. 2019. DLIME: A Deterministic Local Interpretable Model-Agnostic Explanations Approach for Computer-Aided Diagnosis Systems. CoRR abs/1906.10263(2019).Google Scholar
- D Zeiler and R Fergus. 2014. Visualizing and understanding convolutional networks. European Conference on Computer Vision(2014), 818–833.Google ScholarCross Ref
- B Zhou, A Khosla, A Lapedriza, A Oliva, and A Torralba. 2016. Learning deep features for discriminative localization. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2016), 2921–2929.Google ScholarCross Ref
Recommendations
Methods and standards for research on explainable artificial intelligence: Lessons from intelligent tutoring systems
AbstractThe DARPA Explainable Artificial Intelligence (AI) (XAI) Program focused on generating explanations for AI programs that use machine learning techniques. This article highlights progress during the DARPA Program (2017‐2021) relative to research ...
Lessons learned in the work on intelligent tutoring systems that apply to system design in Explainable AI. image image
Explainable health prediction from facial features with transfer learning
In the recent years, Artificial Intelligence (AI) has been widely deployed in the healthcare industry. The new AI technology enables efficient and personalized healthcare systems for the public. In this paper, transfer learning with pre-trained VGGFace ...
Explainable AI and Interpretable Machine Learning: A Case Study in Perspective
AbstractExplainable AI, as the word implies is a type of artificial intelligence which enables the explanation of learning models and focuses on why the system arrived at a particular decision, exploring its logical paradigms, contrary to the inherent ...
Comments