Skip to main content

Sensor fusion for mine detection with the RNN

  • Part VI: Speech, Vision, and Pattern Recognition
  • Conference paper
  • First Online:
Artificial Neural Networks — ICANN'97 (ICANN 1997)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 1327))

Included in the following conference series:

Abstract

In this paper we propose a neural network based approach to sensor fusion, to detect mine locations from electromagnetic induction (EMI) data. Our results use the Random Neural Network (RNN) model [2, 4, 5] which is closer to biophysical reality and mathematically more tractable than standard neural methods. The network is trained to produce an error minimizing non-linear mapping from three sensor output images to the fused image. The result is thresholded to point to likely mine locations.

This research is support by the MURI on Demining, under U.S. Army Research Office Grant No. DAAH04-96-1-0448.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Pearson, J.C., Gelfand, J.J., Sullivan, W.E., Peterson, R.M., Spence, C.D. “Neural network approach to sensory fusion”, Proceedings of SPIE, vol. 931, Sensor Fusion, 1988.

    Google Scholar 

  2. Gelenbe, E. “Random neural networks with negative and positive signals and product form solution”, Neural Computation, vol. 1, no. 4, pp. 502–511, 1989.

    Google Scholar 

  3. Terzopoulos, D. “Physically-based fusion of visual data over space, time, and scale”, Proceedings of the NATA Advanced Research Workshop on Multisensor Fusion for Computer Vision, pp. 63–69, 1989.

    Google Scholar 

  4. Gelenbe, E. “Stability of the random neural network model”, Neural Computation, vol. 2, no. 2, pp. 239–247, 1990.

    Google Scholar 

  5. Gelenbe, E. “Learning in the recurrent random neural network”, Neural Computation, vol. 5, no. 1, pp. 154–164, 1993.

    Google Scholar 

  6. Gelli K., McLauchlan, R.A., Omar, S.I., Challoo, R. “Multisensor fusion/integration using an unsupervised hybrid neural network”, Intelligent Engineering Systems Through Artificial Neural Networks, Vol. 4, ed. Dagli et al. 1994

    Google Scholar 

  7. Hutsberger, T. “Biologically motivated cross-modality sensory fusion system for automatic target recognition”, Neural Networks vol. 8, no.7–8 p.1215–26, 1995.

    Google Scholar 

  8. Goerge, V., Atshuler, T. et al. “Background data collection test plan”, Technical Report, Walcoff Associates, December 1996.

    Google Scholar 

  9. Constantini, M., Farina, A., Zirilli, F. “The fusion of different resolution SAR images”, Proceedings of the IEEE, vol. 85, no. 1, Jan. 1997.

    Google Scholar 

  10. Gelenbe, E., Koçak, T. “Image fusion with the Random Neural Network model”, submitted for publication.

    Google Scholar 

  11. Bakircioglu, H., Gelenbe, E., Koçak, T. “Image processing with the Random Neural Network model”, Proc. 13th International Conference on Digital Signal Processing (DSP97), Santorini, Greece, July 1997.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Wulfram Gerstner Alain Germond Martin Hasler Jean-Daniel Nicoud

Rights and permissions

Reprints and permissions

Copyright information

© 1997 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Gelenbe, E., Koçak, T., Collins, L. (1997). Sensor fusion for mine detection with the RNN. In: Gerstner, W., Germond, A., Hasler, M., Nicoud, JD. (eds) Artificial Neural Networks — ICANN'97. ICANN 1997. Lecture Notes in Computer Science, vol 1327. Springer, Berlin, Heidelberg. https://doi.org/10.1007/BFb0020273

Download citation

  • DOI: https://doi.org/10.1007/BFb0020273

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-63631-1

  • Online ISBN: 978-3-540-69620-9

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics