Skip to main content

Training Set Selection in Neural Network Applications

  • Conference paper

Abstract

In some applications of ANNs to classification problems, training can be inefficient simply because of the high volume of data available for training purposes. The question arises whether it is necessary to use all the data in order adequately to approximate the decision boundaries.

It is intuitively obvious that points near to the boundaries are likely to have more influence over the network weights than those which are far away. Of course, as the boundaries are initially unknown, the status of any particular data point is also unknown. Here we describe an approach which uses an initial crude estimate of the decision boundaries to select appropriate training data in the case of the Multi-Layer Perceptron, followed by a phased addition of points to the training set. We compare this approach with the standard method on both artificial and real data sets, and report results which demonstrates the potential for improved performance in terms of both efficiency and reliability.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. C.R. Reeves (1995) Bias estimation for neural network predictions. This volume.

    Google Scholar 

  2. E.C. Uberbacher and R.J. Mural (1991) Locating protein-coding regions in human DNA sequences by a multiple sensor neural network approach. Proc.Nat.Acad.Sci. USA, 88, 11261–11265.

    Article  Google Scholar 

  3. P. Herdman (1994) Report in Networks: the Official Newsletter of the Neural Computing Applications Forum, November 1994.

    Google Scholar 

  4. M. Wann, T. Hediger and N.N. Greenbaum (1990) The influence of training sets on generalization in feed-forward neural networks. Proc. International Joint Conference on Neural Networks, Vol III, 137–142.

    Article  Google Scholar 

  5. M. Plutowski (1994) Selecting Training Exemplars for Neural Network Learning. PhD Dissertation, University of California, San Diego.

    Google Scholar 

  6. A. Röbel (1994) The Dynamic Pattern Selection Algorithm: Effective Training and Controlled Generalization of Backpropagation Neural Networks. Technical Report, Technical University of Berlin.

    Google Scholar 

  7. W.H. Wolberg and O.L. Mangasarian (1990) Multisurface method of pattern separation for medical diagnosis applied to breast cytology. Proc.Nat. Acad.Sci. USA, 87, 9193–9196.

    Article  MATH  Google Scholar 

  8. C.R. Reeves and N.C. Steele (1993) Neural networks for multivariate analysis: results of some cross-validation studies. Proc. of 6 th International Symposium on Applied Stochastic Models and Data Analysis, World Scientific Publishing, Singapore, Vol II, 780–791.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

Copyright information

© 1995 Springer-Verlag/Wien

About this paper

Cite this paper

Reeves, C.R. (1995). Training Set Selection in Neural Network Applications. In: Artificial Neural Nets and Genetic Algorithms. Springer, Vienna. https://doi.org/10.1007/978-3-7091-7535-4_123

Download citation

  • DOI: https://doi.org/10.1007/978-3-7091-7535-4_123

  • Publisher Name: Springer, Vienna

  • Print ISBN: 978-3-211-82692-8

  • Online ISBN: 978-3-7091-7535-4

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics