Skip to main content

Refitting Solutions Promoted by \(\ell _{12}\) Sparse Analysis Regularizations with Block Penalties

  • Conference paper
  • First Online:
Scale Space and Variational Methods in Computer Vision (SSVM 2019)

Abstract

In inverse problems, the use of an \(\ell _{12}\) analysis regularizer induces a bias in the estimated solution. We propose a general refitting framework for removing this artifact while keeping information of interest contained in the biased solution. This is done through the use of refitting block penalties that only act on the co-support of the estimation. Based on an analysis of related works in the literature, we propose a new penalty that is well suited for refitting purposes. We also present an efficient algorithmic method to obtain the refitted solution along with the original (biased) solution for any convex refitting block penalty. Experiments illustrate the good behavior of the proposed block penalty for refitting.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    As in [2], extended support \(\Vert \hat{z}_i \Vert =\lambda \) can be tackled by testing \(\Vert (\hat{z}^k+\sigma \varGamma \hat{v}^k )_i \Vert \geqslant \lambda \).

References

  1. Belloni, A., Chernozhukov, V.: Least squares after model selection in high-dimensional sparse models. Bernoulli 19(2), 521–547 (2013)

    Article  MathSciNet  Google Scholar 

  2. Brinkmann, E.-M., Burger, M., Rasch, J., Sutour, C.: Bias-reduction in variational regularization. J. Math. Imaging Vis. (2017)

    Google Scholar 

  3. Burger, M., Gilboa, G., Moeller, M., Eckardt, L., Cremers, D.: Spectral decompositions using one-homogeneous functionals. SIAM J. Imaging Sci. 9(3), 1374–1408 (2016)

    Article  MathSciNet  Google Scholar 

  4. Chambolle, A., Pock, T.: A first-order primal-dual algorithm for convex problems with applications to imaging. J. Math. Imaging Vis. 40, 120–145 (2011)

    Article  MathSciNet  Google Scholar 

  5. Chzhen, E., Hebiri, M., Salmon, J.: On lasso refitting strategies. Bernouilli (2019)

    Google Scholar 

  6. Deledalle, C.-A., Papadakis, N., Salmon, J.: On debiasing restoration algorithms: applications to total-variation and nonlocal-means. In: Aujol, J.-F., Nikolova, M., Papadakis, N. (eds.) SSVM 2015. LNCS, vol. 9087, pp. 129–141. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-18461-6_11

    Chapter  Google Scholar 

  7. Deledalle, C.-A., Papadakis, N., Salmon, J., Vaiter, S.: CLEAR: covariant least-square re-fitting with applications to image restoration. SIAM J. Imaging Sci. 10(1), 243–284 (2017)

    Article  MathSciNet  Google Scholar 

  8. Efron, B., Hastie, T., Johnstone, I., Tibshirani, R.: Least angle regression. Ann. Stat. 32(2), 407–499 (2004)

    Article  MathSciNet  Google Scholar 

  9. Lederer, J.: Trust, but verify: benefits and pitfalls of least-squares refitting in high dimensions. arXiv preprint arXiv:1306.0113 (2013)

  10. Osher, S., Burger, M., Goldfarb, D., Xu, J., Yin, W.: An iterative regularization method for total variation-based image restoration. Multiscale Model. Simul. 4(2), 460–489 (2005)

    Article  MathSciNet  Google Scholar 

  11. Rigollet, P., Tsybakov, A.B.: Exponential screening and optimal rates of sparse estimation. Ann. Stat. 39(2), 731–771 (2011)

    Article  MathSciNet  Google Scholar 

  12. Scherzer, O., Groetsch, C.: Inverse scale space theory for inverse problems. In: Kerckhove, M. (ed.) Scale-Space 2001. LNCS, vol. 2106, pp. 317–325. Springer, Heidelberg (2001). https://doi.org/10.1007/3-540-47778-0_29

    Chapter  MATH  Google Scholar 

  13. Tadmor, E., Nezzar, S., Vese, L.: A multiscale image representation using hierarchical (BV, L2) decompositions. Multiscale Model. Simul. 2(4), 554–579 (2004)

    Article  MathSciNet  Google Scholar 

Download references

Acknowledgement

This project has been carried out with support from the French State, managed by the French National Research Agency (ANR-16-CE33-0010-01). This work was supported by the European Union’s Horizon 2020 research and innovation programme under the Marie Skłodowska-Curie grant agreement No 777826.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Nicolas Papadakis .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Deledalle, CA., Papadakis, N., Salmon, J., Vaiter, S. (2019). Refitting Solutions Promoted by \(\ell _{12}\) Sparse Analysis Regularizations with Block Penalties. In: Lellmann, J., Burger, M., Modersitzki, J. (eds) Scale Space and Variational Methods in Computer Vision. SSVM 2019. Lecture Notes in Computer Science(), vol 11603. Springer, Cham. https://doi.org/10.1007/978-3-030-22368-7_11

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-22368-7_11

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-22367-0

  • Online ISBN: 978-3-030-22368-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics