Review of advances in neural networks: Neural design technology stack
Section snippets
Introduction and background
Recently, we have been witnessing increasing interest in understanding how the brain functions and how it could be modeled. This momentum is fueled by major research initiatives, including the Human Brain Project [44], the BRAIN Initiative, as well as significant commercial efforts, such as IBM Watson [20].
In this paper we provide a high-level overview of the state-of-the-art in neural networks, the main building block of the brain. Computer scientists typically focus on artificial neural
Neural network definitions
From the structural point of view, a brain consists of a network of neurons, densely interconnected via synapses – a structure which artificial neural networks are attempting to recreate. In this section we present different neuron and network models, as well as discuss the information and feature encoding in these networks.
Establishing model parameters
A neural network is a meta-model for computation. Applying it to solve a concrete problem requires a particular set of parameters, whose tuning and life cycle are presented in the following subsections.
Learning
Learning is a data-driven mechanism for obtaining model parameters. In this section, we describe the learning process and we make a classification of learning algorithms, with a focus on a number of interesting learning trends.
Machine learning techniques aim to shift the work burden from human effort to machine computation, for specific classes of problems. With the right algorithm, the manual effort of designing solutions can be traded off for memory and computational power. Algorithms, in the
Advancing neural networks research
A number of recent advancements have helped neural networks to regain mainstream attention. Firstly, novel hardware has made it possible to verify concepts at a much larger scale, e.g. by simulating a system with 109 neurons [2], comparable in scale to a cat׳s brain. Secondly, the unsupervised feature learning paradigm offered a logic that could utilize large unpreprocessed datasets, providing results such as a self-emerging cat image detector [43]. Having already discussed the latter in
Summary
In this paper, we have provided a high-level overview of current trends in neural network research. By starting from the definitions of neural network models, we compare the artificial neuron model with the biologically-plausible spiking neuron. Furthermore, we explain the basic concepts behind the encoding and decoding of information in a neural network model, be it an artificial mathematical abstraction, or a biological neuroscience model. Moreover, we present the model parameters that need
Adela-Diana Almási obtained her Bachelor׳s and Master׳s degrees in Computer Science from the “Politehnica” University of Bucharest, Romania, where she is currently a PhD student under the supervision of Prof. Valentin Cristea. She is conducting her doctoral research in collaboration with the IBM Research Laboratory in Zurich, Switzerland. Her main research interests are machine learning, neural networks and data analytics.
References (68)
- et al.
A dynamic spatial gradient of Hebbian learning in dendrites
Neuron
(2006) - et al.
Error-backpropagation in temporally encoded networks of spiking neurons
Neurocomputing
(2002) - et al.
Synchronous oscillatory neural ensembles for rules in the prefrontal cortex
Neuron
(2012) - et al.
Chaotic simulated annealing by a neural network model with transient chaos
Neural Netw.
(1995) - et al.
Multi-column deep neural network for traffic sign classification
Neural Netw.
(2012) - et al.
Face processing using one spike per neurone
Biosystems
(1998) - et al.
The impulses produced by sensory nerve-endings. Part II: the response of a single end-organ
J. Physiol.
(1926) - R. Ananthanarayanan, S.K. Esser, H.D. Simon, D.S. Modha, The cat is out of the bag, In: Proceedings of the Conference...
- et al.
Information theory and neural coding
Nat. Neurosci.
(1999) - et al.
A comprehensive workflow for general-purpose neural modeling with highly configurable neuromorphic hardware systems
Biol. Cybern.
(2011)
Context-dependent pre-trained deep neural networks for large-vocabulary speech recognition
IEEE Trans. Audio Speech Lang. Process.
Theoretical neuroscience: computational and mathematical modeling of neural systems
J. Cognit. Neurosci.
Axon physiology
Physiol. Rev.
How to Build a BrainA Neural Architecture for Biological Cognition
Neural EngineeringComputation, Representation, and Dynamics in Neurobiological Systems
A large-scale model of the functioning brain
Science
Designing neural networks using gene expression programming
Appl. Soft Comput. Technol.: Chall. Complex.
Building Watsonan overview of the deepqa Project
AI Mag.
Functional requirements for reward-modulated spike-timing-dependent plasticity
J. Neurosci.
Combining ELMs with random projections
IEEE Intell. Syst.
A neuronal learning rule for sub-millisecond temporal coding
Nature
Spiking Neuron ModelsSingle Neurons, Populations, Plasticity
How good are neuron models?
Science
Theory and simulation in neuroscience
Science
A novel connectionist system for unconstrained handwriting recognition
IEEE Trans. Pattern Anal. Mach. Intell.
Evolutionary neural networks for anomaly detection based on the behavior of a program
IEEE Trans. Syst. Man Cybern. Part B: Cybern.
ModelDBa database to support computational neuroscience
J. Comput. Neurosci.
A fast learning algorithm for deep belief nets
Neural Comput.
Cited by (64)
Challenges in implementing data-driven approaches for building life cycle energy assessment: A review
2022, Renewable and Sustainable Energy ReviewsDeep learning for predictions of hydrolysis rates and conditional molecular design of esters
2021, Journal of the Taiwan Institute of Chemical EngineersTask-Adaptive Neuromorphic Computing Using Reconfigurable Organic Neuristors with Tunable Plasticity and Logic-in-Memory Operations
2024, Journal of Physical Chemistry LettersDiagnostic reasoning in the age of artificial intelligence: Synergy or opposition?
2024, Journal of Hospital Medicine
Adela-Diana Almási obtained her Bachelor׳s and Master׳s degrees in Computer Science from the “Politehnica” University of Bucharest, Romania, where she is currently a PhD student under the supervision of Prof. Valentin Cristea. She is conducting her doctoral research in collaboration with the IBM Research Laboratory in Zurich, Switzerland. Her main research interests are machine learning, neural networks and data analytics.
Stanisław Woźniak received his BSc and MSc degrees in Computer Science from Poznan University of Technology, Poland. Currently he is pursuing a PhD degree at École Polytechnique Fédérale de Lausanne, Switzerland, in collaboration with IBM Research – Zurich, Switzerland. His research interests include neural networks, machine learning, computational neuroscience and cognitive systems.
Valentin Cristea is the Head of the Computer Science and Engineering Department and Professor at the Politehnica University of Bucharest. His main fields of expertise are Computer networks, E-Services, and Large Scale Distributed Systems, topics on which he teaches courses and supervises PhD students. He is the Director of the National Center for Information Technology, leader of the CoLaborator, Distributed Systems and Grid, and e-Business/e-Government laboratories. He has a long experience in the development and management of international and national research projects on e-services, dependable large scale distributed systems, Grid and Cloud computing, and smart environments. He received the IBM Faculty Award in 2003 and 2011, he is an IEEE and ACM member and a Phare IT expert.
Yusuf Leblebici received the B.Sc. and M.Sc. degrees in electrical engineering from Istanbul Technical University, Istanbul, Turkey, in 1984 and 1986, respectively, and the Ph.D. degree in electrical and computer engineering from the University of Illinois, Urbana-Champaign (UIUC), in 1990. Since 2002, he is a Chair Professor at the Swiss Federal Institute of Technology in Lausanne (EPFL), and director of Microelectronic Systems Laboratory. His research interests include design of high-speed CMOS digital and mixed-signal integrated circuits, computer-aided design of VLSI systems, intelligent sensor interfaces, modeling and simulation of semiconductor devices, and VLSI reliability analysis. He is the coauthor of six textbooks, as well as more than 300 articles published in various journals and conferences. He is a Fellow of IEEE since 2010, and he has been elected as Distinguished Lecturer of the IEEE Circuits and Systems Society for 2010–2011.
Ton Engbersen has been with the IBM Research since 1980. His career spanned such diverse areas as Image processing, chip design, communications technology, server technology, legacy management, Innovation in Outsourcing and Data center Energy management. Throughout the years he has held a range of management positions in R&D in Switzerland and in the US. As member of the IBM Academy of Technology he led the European branch from 2009 to 2011. Ton has published more than 50 articles in refereed scientific journals. Currently he is the Scientific Director for the ASTRON-IBM Center of Exascale Technology, and leading the DOME Project http://www-03.ibm.com/press/us/en/pressrelease/37361.wss.