Abstract
There have been two competing paradigms in artificial intelligence (AI) development ever since its birth in 1956, i.e., symbolism and connectionism (or sub-symbolism). While symbolism dominated AI research by the end of 1980s, connectionism gained momentum in the 1990s and is gradually displacing symbolism. This paper considers symbolism as the first generation of AI and connectionism as the second generation. However, each of these two paradigms simulates the human mind from only one perspective. AI cannot achieve true human behaviors by relying on only one paradigm. In order to develop novel AI technologies that are safe, reliable, and extensible, it is necessary to establish a new explainable and robust AI theory. To this end, this paper looks toward developing a third generation artificial intelligence by combining the current paradigms.
Similar content being viewed by others
References
Simon H A. Models of Man. New York: Wiley & Sons, 1957
Newell A, Simon H A. Computer science as empirical inquiry: symbols and search. Commun ACM, 1976, 19: 113–126
Newell A. Physical Symbol Systems. Cogn Sci, 1980, 4: 135–183
Fodor J A. Methodological solipsism considered as a research strategy in cognitive psychology. Behav Brain Sci, 1980, 3: 63–73
McCarthy J, Minsky M L, Rochester N, et al. A proposal for the Dartmouth summer research project on artificial intelligence. 1955, 27: 4
Lindsay, Robert K, Bruce G. Applications of Artificial Intelligence for Organic Chemistry: The Dendral Project. New York: McGraw-Hill Book Company, 1980
Buchanan B G, Shortliffe E H. Rule-Based Expert Systems: The MYCIN Experiments of the Stanford Heuristic Programming Project. Boston: Addison Wesley, 1984
Muggleton S, de Raedt L. Inductive logic programming: theory and methods. J Logic Programm, 1994, 19–20: 629–679
Riguzzi F, Bellodi E, Zese R. A history of probabilistic inductive logic programming. Front Robot AI, 2014, 1: 6
Yang Q, Zhang Y, Dai W Y, et al. Transfer Learning. Cambridge: Cambridge University Press, 2020
Ehrlinger L, Wolfram W. Towards a definition of knowledge graphs. In: Joint Proceedings of the Posters and Demos Track of 12th International Conference on Semantic Systems (SEMANTiCS2016) and 1st International Workshop on Semantic Change & Evolving Semantics (SuCCESS16), Leipzig, 2016
Russel S J, Norvig P. Artificial Intelligence: A Modern Approach. 2nd ed. New York: Pearson Education Inc., 2003
Rosenblatt F. The perceptron: a probabilistic model for information storage and organization in the brain. Psychol Rev, 1958, 65: 6
Rosenblatt F. Principles of Neurodynamics. Arch Gen Psychiatry, 1962 7: 218–219
McCulloch W S, Pitts W. A logical calculus of the ideas immanent in nervous activity. Bull Math Biophys, 1943, 5: 115–133
Hebb D O. The Organization of Behavior: A Neuropsychological Theory. London: Psychology Press, 1949
Minsky M, Papert S A. Perceptrons: An Introduction to Computational Geometry. Cambridge: MIT Press, 1969
Cauchy A. Methode generale pour la resolution des systemes d’equations simultanees. Comp Rend Acad Sci, Paris, 1847, 25: 536–538
Nesterov Y E. A method for solving the convex programming problem with convergence rate O(1/k2). Soviet Math Doklady, 1983, 27: 372–376
Linnainmaa S. The representation of the cumulative rounding error of an algorithm as a taylor expansion of the local rounding errors (in Finnish). Dissertation for Master’s Degree. Helsinki: University of Helsinki, 1970
Rumelhart D E, Hinton G E, Williams R J. Learning representations by back-propagating errors. Nature, 1986, 323: 533–536
Janocha K, Czarnecki W M. On loss functions for deep neural networks in classification. Schedae Inform, 2016, 25: 49–59
Wan L, Zeiler M, Zhang A X, et al. Regularization of neural networks using DropConnect. In: Proceedings of the 30th International Conference on Machine Learning (ICML), Atlanta, 2013
Fukushima K. Neocognitron: a self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position. Biol Cybernetics, 1980, 36: 193–202
Derevyanko G, Grudinin S, Bengio Y, et al. Deep convolutional networks for quality assessment of protein folds. Bioinformatics, 2018, 34: 4046–4053
Lipton J C, Berkowitz J. A critical review of recurrent neural networks for sequence learning. 2015. ArXiv:1506.00019v4
Hochreiter S, Schmidhuber J. Long short-term memory. Neural Computation, 1997, 9: 1735–1780
Hinton G E, Osindero S, Teh Y W. A fast learning algorithm for deep belief nets. Neural Comput, 2006, 18: 1527–1554
LeCun Y, Bengio Y, Hinton G. Deep learning. Nature, 2015, 521: 436–444
Coulom R. Efficient selectivity and backup operators in Monte-Carlo tree search. In: Proceedings of International Conference on Computers and Games, Berlin, 2006
Sutton R S, Barto A G. Reinforcement Learning: An Introduction. Cambridge: MIT Press, 2015
Dong Y P, Liao F Z, Pang T Y, et al. Boosting adversarial attacks with momentum. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, 2018
Dong Y, Su H, Zhu J, et al. Towards interpretable deep neural networks by leveraging adversarial examples. In: Proceedings of IJCAI Workshop on AISC, Sydney, 2019
Ferrucci D, Brown E, Chu-Carroll J, et al. Building Watson: an overview of the DeepQA project. In: Proceedings of Association for the Advancement of Artificial Intelligence (AAAI), Atlanta, 2010
Harnad S. The symbol grounding problem. Phys D-Nonlin Phenom, 1990, 42: 335–346
Chen X, Duan Y, Houthooft R, et al. InfoGAN: interpretable representation learning by information maximizing generative adversarial nets. In: Proceedings of Advances in Neural Information Processing Systems (NeurIPS), Barcelona, 2016. 2172–2180
Liu Y, Wei F Y, Shao J, et al. Exploring disentangled feature representation beyond face identification. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition(CVPR), Salt Lake City, 2018. 2080–2089
Higgins I, Matthey L, Pal A, et al. Beta-VAE: learning basic visual concepts with a constrained variational framework. In: Proceedings of the 32nd International Conference on Logic Programming (ICLP), New York City, 2016
Siddharth N, Paige B, Desmaison A, et al. Inducing interpretable representations with variational autoencoders. In: Proceedings of Advances in Neural Information Processing Systems (NeurIPS), Barcelona, 2016
Li C, Xu K, Zhu J, et al. Triple generative adversarial nets. In: Proceedings of Advances in Neural Information Processing Systems (NeurIPS), Long Beach, 2017
Hu X L, Zhang J W, Li J M, et al. Sparsity regularized HMAX for visual recognition. Plos One, 2014, 9: e81813
Silver D, Singh S, Precup D, et al. Reward is enough. Artif Intell, 2021, 299: 103535
Mnih V, Kavukcuoglu K, Silver D, et al. Human-level control through deep reinforcement learning. Nature, 2015, 518: 529–533
Vinyals O, Babuschkin I, Czarnecki W M, et al. Grandmaster level in StarCraft II using multi-agent reinforcement learning. Nature, 2019, 575: 350–354
Silver D, Huang A, Maddison C J, et al. Mastering the game of Go with deep neural networks and tree search. Nature, 2016, 529: 484–489
Silver D, Schrittwieser J, Simonyan K, et al. Mastering the game of Go without human knowledge. Nature, 2017, 550: 354–359
Duan Y, Chen X, Houthooft R, et al. Benchmarking deep reinforcement learning for continuous control. In: Proceedings of International Conference on Machine Learning (ICML), New York, 2016
Schulman J, Wolski F, Dhariwal P, et al. Proximal policy optimization algorithms. 2017. ArXiv:1707.06347
Sutton R S, Barto A G. Reinforcement Learning: An Introduction. 2nd ed. Cambridge: MIT Press, 2015
François-Lavet V, Henderson P, Islam R, et al. An introduction to deep reinforcement learning. FNT Machine Learn, 2018, 11: 219–354
Huang S Y, Su H, Zhu J, et al. SVQN: sequential variational soft Q-learning networks. In: Proceedings of International Conference on Learning Representations (ICLR), 2020
Huang S Y, Su H, Zhu J, et al. Combo-action: training agent for FPS game with auxiliary tasks. In: Proceedings of the 33rd AAAI Conference on Artificial Intelligence (AAAI), Honolulu, 2019
Song S H, Weng J Y, Su H, et al. Playing FPS games with environment-aware hierarchical reinforcement learning. In: Proceedings of the 28th International Joint Conference on Artificial Intelligence (IJCAI), Macau, 2019
Zhou Y C, Li J L, Zhu J. Posterior sampling for multi-agent reinforcement learning: solving extensive games with imperfect information. In: Proceedings of International Conference on Learning Representations (ICLR), Addis Ababa, 2020
Mikolov T, Chen K, Corrado G, et al. Efficient estimation of word representations in vector space. In: Proceedings of International Conference on Learning Representations (ICLR), Scottsdale, 2013
Pennington J, Socher R, Manning C D. Global vectors for word representation. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, 2014
Mikolov T, Sutskever I, Chen K, et al. Distributed representations of words and phrases and their compositionality. In: Proceedings of Advances in Neural Information Processing Systems (NeurIPS), Lake Tahoe, 2013
Kusner M J, Sun Y, Kolkin N I, et al. From word embeddings to document distances. In: Proceedings of the 32nd International Conference on Machine Learning (ICML), Lille, 2015
Zhang J, Liu Y, Luan H B, et al. Prior knowledge integration for neural machine translation using posterior regularization. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (ACL), 2017
Ding Y, Liu Y, Luan H B, et al, Visualizing and understanding neural machine translation. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (ACL), 2017
Zhang B. Artificial intelligence in the post-deep learning era (in Chinese). CAAI Trans Intell Technol, 2017, 7: 3–5
Liu M, Shi J, Li Z, et al. Towards better analysis of deep convolutional neural networks. In: Proceedings of IEEE Conference on Visual Analytics Science and Technology, 2016
Wang Y L, Su H, Hu X L. Interpret neural networks by identifying critical data routing paths. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, 2018
Hinton G, Vinyals O, Dean J. Distilling the knowledge in a neural network. In: Proceedings of Advances in Neural Information Processing Systems (NeurIPS), Montreal, 2014
Dong Y P, Su H, Zhu J, et al. Improving interpretability of deep neural networks with semantic information. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, 2017
Dong Y P, Fu Q-A, Yang X, et al. Benchmarking adversarial robustness on image classification. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, 2020
Dziugaite G K, Ghahramani Z, Roy D M. A study of the effect of jpg compression on adversarial images. 2016. ArXiv:1608.00853
Samangouei P, Kabkab M, Chellappa R. Defense-GAN: protecting classifiers against adversarial attacks using generative models. In: Proceedings of International Conference on Learning Representations (ICLR), 2018
Liao F Z, Liang M, Dong Y P, et al. Defense against adversarial attacks using high-level representation guided denoiser. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, 2018
Madry A, Makelov A, Schmidt L, et al. Towards deep learning models resistant to adversarial attacks. In: Proceedings of International Conference on Learning Representations (ICLR), 2018
Tramer F, Carlini N, Brendel W, et al. On adaptive attacks to adversarial example defenses. 2020. ArXiv:2002.08347
Dong Y P, Deng Z J, Pang T Y, et al. Adversarial distributional training for robust deep learning. In: Proceedings of Advances in Neural Information Processing Systems (NeurIPS), 2020
Kurakin A, Goodfellow I, Bengio S, et al. Adversarial attacks and defences competition. 2018. ArXiv:1804.00097
Liu X Q, Cheng M H, Zhang H, et al. Towards robust neural networks via random selfensemble. In: Proceedings of the European Conference on Computer Vision (ECCV), 2018
Pang T Y, Xu K, Du C, et al. Improving adversarial robustness via promoting ensemble diversity. In: Proceedings of International Conference on Machine Learning (ICML), Long Beach, 2019
Castelvecchi D. Can we open the black box of AI? Nature, 2016, 538: 20–23
Ghahramani Z. Probabilistic machine learning and artificial intelligence. Nature, 2015, 521: 452–459
Zhu J, Chen J, Hu W, et al. Big learning with Bayesian methods. Natl Sci Rev, 2017, 4: 627–651
Zhu J, Chen N, Xing E P. Bayesian inference with posterior regularization and applications to infinite latent SVMs. J Mach Learn Res, 2014, 15: 1799–1847
Williams P M. Bayesian conditionalisation and the principle of minimum information. Br J Philosophy Sci, 1980, 31: 131–144
Mei S, Zhu J, Zhu X. Robust RegBayes: selectively incorporating first-order logic domain knowledge into Bayesian models. In: Proceedings of International Conference on Machine Learning (ICML), Beijing, 2014
Zhu J, Ahmed A, Xing E P. MedLDA: maximum margin supervised topic models. J Mach Learn Res, 2012, 13: 2237–2278
MacKay D J C. Bayesian methods for adaptive models. Dissertation for Ph.D. Degree. Pasadena: California Institute of Technology, 1992
Neal R M. Bayesian learning for neural networks. Dissertation for Ph.D. Degree. Toronto: University of Toronto, 1995
Shi J, Sun S, Zhu J. A spectral approach to gradient estimation for implicit distributions. In: Proceedings of the 35th International Conference on Machine Learning (ICML), Stockholm, 2018
Zhou Y, Shi J, Zhu J. Nonparametric score estimators. In: Proceedings of International Conference on Machine Learning (ICML), Vienna, 2020
Wang Z, Ren T, Zhu J, et al. Function space particle optimization for Bayesian neural networks. In: Proceedings of International Conference on Learning Representations (ICLR), New Orleans, 2019
Xu K, Li C, Zhu J, et al. Understanding and stabilizing GANs’ training dynamics using control theory. In: Proceedings of International Conference on Machine Learning (ICML), Vienna, 2020
Chen J, Lu C, Chenli B, et al. VFlow: more expressive generative flows with variational data augmentation. In: Proceedings of International Conference on Machine Learning (ICML), Vienna, 2020
Shi J, Chen J, Zhu J, et al. ZhuSuan: a library for Bayesian deep learning. 2017. ArXiv:1709.05870
Li J Z, Su H, Zhu J, et al. Essay-anchor attentive multi-modal bilinear pooling for textbook question answering. In: Proceedings of IEEE International Conference on Multimedia and Expo (ICME), San Diego, 2018
Ren M, Kiros R, Zemel R S. Exploring models and data for image question answering. In: Proceedings of Advances in Neural Information Processing Systems (NeurIPS), 2015
Zhu Y, Groth O, Bernstein M S, et al. Visual7W: grounded question answering in images. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, 2016
Acknowledgements
This work was supported by National Natural Science Foundation of China (Grant No. 61620106010).
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Zhang, B., Zhu, J. & Su, H. Toward the third generation artificial intelligence. Sci. China Inf. Sci. 66, 121101 (2023). https://doi.org/10.1007/s11432-021-3449-x
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/s11432-021-3449-x