A hybrid neural approach to combinatorial optimization

https://doi.org/10.1016/0305-0548(95)00064-XGet rights and content

Abstract

Both the Hopfield neural network and Kohonen's principles of self-organization have been used to solve difficult optimization problems, with varying degrees of success. In this paper, a hybrid neural network is presented which combines, for the first time, a new self-organizing approach to optimization with a Hopfield network. It is demonstrated that many of the traditional problems associated with each of these approaches can be resolved when they are combined into a hybrid model. After presenting the broad class of 0–1 optimization problems for which this hybrid neural approach is suited, details of the algorithm are presented and convergence properties are discussed. This hybrid neural approach is applied to solve a practical sequencing problem from the car manufacturing industry. Performance results are compared with classical as well as other neural techniques, and conclusions are drawn.

References (28)

  • B. Angéniol et al.

    Self organizing feature maps and the Travelling Salesman Problem

    Neural Networks

    (1988)
  • J.J. Hopfield et al.

    “Neural” computation of decisions in optimization problems

    Biological Cybernet.

    (1985)
  • D.W. Tank et al.

    Simple neural optimization networks: an A/D converter, signal decision circuit and a linear programming circuit

    IEEE Trans. Circuit Syst.

    (1986)
  • G. Tagliarini et al.

    Solving Constraint Satisfaction Problems with neural networks

  • G.V. Reklaitis et al.

    Generalised Hopfield Networks and nonlinear optimization

  • G.V. Wilson et al.

    On the stability of the TSP algorithm of Hopfield and Tank

    Biological Cybernet.

    (1988)
  • S. Hedge et al.

    Determination of parameters in a Hopfield/Tank computational network

  • B. Kamgar-Parsi et al.

    Dynamical stability and parameter selection in neural optimization

  • W.K. Lai et al.

    Genetic breeding of control parameters for the Hopfield.Tank neural net

  • R.D. Brandt et al.

    Alternative networks for solving the Travelling Salesman Problem and the List-Matching Problem

  • D.E. Van den Bout et al.

    A Travelling Salesman objective function that works

  • A.H. Gee et al.

    Limitations of neural networks for solving Traveling Salesman Problems

    IEEE Trans. Neural Networks

    (1995)
  • T. Kohonen

    Self-organized formation of topologically correct feature maps

    Biological Cybernet.

    (1982)
  • L. Fang et al.

    Design of competition-based neural networks for combinatorial optimization

    Int. J. Neural Syst.

    (1990)
  • Cited by (42)

    • A new hybrid quadratic regression and cascade forward backpropagation neural network

      2016, Neurocomputing
      Citation Excerpt :

      The entirety of these investigations advocated that hybrid methods had improved forecasting performance equated to individual forecasting methods. Wolport [8], Smith et al., [9] as well as Zhang and Beradi [10] were some of the earliest authors to propose hybrid methods by merging different neural network architectures and showed that hybrid methods involving different neural network architecture produced better forecasting performance compared to individual neural network methods. In our work, new trends in hybrid methods involving the combination of classical statistical methods such as quadratic regression methods with different neural network architectures such as the cascade forward back propagation neural network are proposed.

    • A neural model for the p-median problem

      2008, Computers and Operations Research
    • Heuristics and augmented neural networks for task scheduling with non-identical machines

      2006, European Journal of Operational Research
      Citation Excerpt :

      This enhancement significantly reduced the inherent complexity of the Hopfield network. Smith et al. (1996) propose a hybrid of Hopfield network and self-organizing networks to overcome the deficiencies of each of these types of network. Smith et al. (1998) explain the Hopfield network, an improved Hopfield network and the self-organizing network for a general optimization problem.

    • Solving terminal assignment problems with groups encoding: The wedding banquet problem

      2006, Engineering Applications of Artificial Intelligence
      Citation Excerpt :

      An important class of combinatorial optimization problems with constraints (Smith et al., 1996; Smith, 1999) can be defined as the maximization or minimization of a goal function (usually in a binary search space) subject to a set of constraints, which have to be satisfied for a solution to be feasible.

    • A portable and scalable algorithm for a class of constrained combinatorial optimization problems

      2005, Computers and Operations Research
      Citation Excerpt :

      In this paper, we refer to this kind of problems as constrained combinatorial optimization problems (CCOPs). Several methods which have been reported as good approaches to concrete examples of CCOPs can be found in the literature [10–14,30,31]. However, these existing algorithms have two main drawbacks: first, the majority of them show lack of scalability.

    View all citing articles on Scopus

    Kate Smith received a Bachelor of Science degree in Mathematics from the University of Melbourne, Australia where she has recently completed Ph.D. studies in the area of neural networks for combinatorial optimization. She is currently a Lecturer in the Department of Business Systems at Monash University, Australia.

    §

    M. Krishnamoorthy is a research scientist with the Operations Research Group at the CSIRO Division of Mathematics and Statistics in Melbourne, Australia. He obtained his M.Sc. and Ph.D. from Imperial College, University of London. He also has an M.Sc. (1985) in Operational Research from the University of Delhi. His research interests are in developing efficient algorithms for combinatorial, network and graph optimization problems.

    View full text