Skip to main content
Log in

Variable selection in uncertain regression analysis with imprecise observations

  • Mathematical methods in data science
  • Published:
Soft Computing Aims and scope Submit manuscript

Abstract

Variable selection is crucial in order to better investigate relationships between variables in regression analysis. However, sometimes data are collected in an imprecise way and can not be described by random variables. As a result, classical variable selection methods are invalid. Characterizing these imprecise observations as uncertain variables, this paper presents the uncertain lasso estimate and the de-biased uncertain lasso estimate to select variables and estimate unknown parameters, respectively. Moreover, a way to choose the tuning parameter using cross-validation is suggested. Finally, numerical examples are documented to show our methods in detail.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Fig. 1
Fig. 2
Fig. 3

Similar content being viewed by others

References

Download references

Acknowledgements

This work was supported by National Natural Science Foundation of China (No. 62073009) and the Program for Young Excellent Talents in UIBE (No. 18YQ06).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xiangfeng Yang.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Ethical approval

This paper does not contain any studies with human participants or animals performed by any of the authors.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix

Appendix

Proof of Corollary 1

According to Definitions 1 and 2, the uncertain lasso estimate and de-biased uncertain lasso estimate for linear regression model (11) solve minimization problems (12) and (13), respectively. Since

$$\begin{aligned} \beta _{0}+\sum _{j=1}^{p} \beta _{j}\tilde{x}_{ji} \end{aligned}$$

is increasing with respect to \(\tilde{x}_{ji}\) when \(\beta _{j} > 0\) and decreasing with respect to \(\tilde{x}_{ji}\) when \(\beta _{j} \le 0\) for each i \((i=1,2,\ldots , n)\), the corollary follows from Theorems 1 and 2 immediately. \(\square \)

Proof of Corollary 2

According to Definitions 1 and 2, the uncertain lasso estimate and de-biased uncertain lasso estimate for the regression model (14) solve minimization problems (15) and (16), respectively. It follows from the operational law of uncertain variable (p. 55 of Liu 2015) that inverse uncertainty distributions of uncertain variables \(\ln \tilde{y}_{i}\) are \(\ln F_{i}^{-1}(\alpha )\), \(i=1,2,\ldots ,n\), respectively. Since

$$\begin{aligned} \beta _{0}+ \sum _{j=1}^{p} \beta _{j} \tilde{x}_{ji} \end{aligned}$$

is increasing with respect to \(\tilde{x}_{ji}\) when \(\beta _{j} > 0\) and decreasing with respect to \(\tilde{x}_{ji}\) when \(\beta _{j} \le 0\) for each i \((i=1,2,\ldots , n)\), the corollary follows from Theorems 1 and 2 immediately.

\(\square \)

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Liu, Z., Yang, X. Variable selection in uncertain regression analysis with imprecise observations. Soft Comput 25, 13377–13387 (2021). https://doi.org/10.1007/s00500-021-06129-x

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00500-021-06129-x

Keywords

Navigation