Skip to main content

Lagrange Programming Neural Networks for Compressive Sampling

  • Conference paper
Neural Information Processing. Models and Applications (ICONIP 2010)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 6444))

Included in the following conference series:

Abstract

Compressive sampling is a sampling technique for sparse signals. The advantage of compressive sampling is that signals are compactly represented by a few number of measured values. This paper adopts an analog neural network technique, Lagrange programming neural networks (LPNNs), to recover data in compressive sampling. We propose the LPNN dynamics to handle three sceneries, including the standard recovery of sparse signal, the recovery of non-sparse signal, and the noisy measurement values, in compressive sampling. Simulation examples demonstrate that our approach effectively recovers the signals from the measured values for both noise free and noisy environment.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Candès, E.J., Wakin, M.B.: An introduction to compressive sampling. In: IEEE Signal Processing, vol. 25, pp. 21–30 (March 2008)

    Google Scholar 

  2. Romberg, J.: Imaging via compressive sampling. In: IEEE Signal Processing, vol. 25, pp. 14–20 (March 2008)

    Google Scholar 

  3. Dennis, J.B.: Mathematical Programming and Electrical Networks. Wiley, New York (1959)

    Google Scholar 

  4. Graf, H., Jackel, L.: Analog electronic neural network circuits. IEEE Circuits Devices Mgag. 5(4), 44–49 (1989)

    Article  Google Scholar 

  5. Mead, C.: Analog VLSI and Neural Systems. Wesley, London (1989)

    MATH  Google Scholar 

  6. Hopfield, J.J.: Neural networks and physical systems with emergent collective computational abilities. Proc. of the National Academy of Sciences 79, 2554–2558 (1982)

    Article  MathSciNet  MATH  Google Scholar 

  7. Chua, L., Lin, G.: Nonlinear programming without computation. IEEE Trans. on Circuits Syst. 31, 182–188 (1984)

    Article  MathSciNet  Google Scholar 

  8. Sum, J., Leung, C.-S., Tam, P., Young, G., Kan, W., Chan, L.-W.: Analysis for a class of winner-take-all model. IEEE Trans. Neural Networks 10(1), 64–71 (1999)

    Article  Google Scholar 

  9. Zhang, S., Constantinidies, A.G.: Lagrange programming neural networks. IEEE Trans. on Circuits and Systems II 39, 441–452 (1992)

    Article  Google Scholar 

  10. Coifman, R., Geshwind, F., Meyer, Y.: Noiselets. Appl. Comp. Harmon. Anal. 10, 27–44 (2001)

    Article  MathSciNet  MATH  Google Scholar 

  11. Kosko, B.: Bidirectional associative memories. IEEE Trans. Syst. Man Cybern. 18(1), 49–60 (1988)

    Article  MathSciNet  Google Scholar 

  12. Leung, C.-S., Chan, L.-W., Lai, E.: Stability and statistical properties of second-order bidirectional associative memory. IEEE Trans. Neural Networks 8(2), 267–277 (1997)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2010 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Lam, PM., Leung, C.S., Sum, J., Constantinides, A.G. (2010). Lagrange Programming Neural Networks for Compressive Sampling. In: Wong, K.W., Mendis, B.S.U., Bouzerdoum, A. (eds) Neural Information Processing. Models and Applications. ICONIP 2010. Lecture Notes in Computer Science, vol 6444. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-17534-3_22

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-17534-3_22

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-17533-6

  • Online ISBN: 978-3-642-17534-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics