IDEAS home Printed from https://ideas.repec.org/a/gam/jmathe/v11y2023i14p3202-d1199385.html
   My bibliography  Save this article

Overview of High-Dimensional Measurement Error Regression Models

Author

Listed:
  • Jingxuan Luo

    (School of Statistics, Beijing Normal University, Beijing 100875, China)

  • Lili Yue

    (School of Statistics and Data Science, Nanjing Audit University, Nanjing 211815, China)

  • Gaorong Li

    (School of Statistics, Beijing Normal University, Beijing 100875, China)

Abstract

High-dimensional measurement error data are becoming more prevalent across various fields. Research on measurement error regression models has gained momentum due to the risk of drawing inaccurate conclusions if measurement errors are ignored. When the dimension p is larger than the sample size n , it is challenging to develop statistical inference methods for high-dimensional measurement error regression models due to the existence of bias, nonconvexity of the objective function, high computational cost and many other difficulties. Over the past few years, some works have overcome the aforementioned difficulties and proposed several novel statistical inference methods. This paper mainly reviews the current development on estimation, hypothesis testing and variable screening methods for high-dimensional measurement error regression models and shows the theoretical results of these methods with some directions worthy of exploring in future research.

Suggested Citation

  • Jingxuan Luo & Lili Yue & Gaorong Li, 2023. "Overview of High-Dimensional Measurement Error Regression Models," Mathematics, MDPI, vol. 11(14), pages 1-22, July.
  • Handle: RePEc:gam:jmathe:v:11:y:2023:i:14:p:3202-:d:1199385
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/2227-7390/11/14/3202/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/2227-7390/11/14/3202/
    Download Restriction: no
    ---><---

    References listed on IDEAS

    as
    1. Faming Liang & Bochao Jia & Jingnan Xue & Qizhai Li & Ye Luo, 2018. "An imputation–regularized optimization algorithm for high dimensional missing data problems and beyond," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 80(5), pages 899-926, November.
    2. Jun Zhang, 2021. "Estimation and variable selection for partial linear single-index distortion measurement errors models," Statistical Papers, Springer, vol. 62(2), pages 887-913, April.
    3. Cui, Hengjian & Chen, Song Xi, 2003. "Empirical likelihood confidence region for parameter in the errors-in-variables models," Journal of Multivariate Analysis, Elsevier, vol. 84(1), pages 101-115, January.
    4. Emre Barut & Jianqing Fan & Anneleen Verhasselt, 2016. "Conditional Sure Independence Screening," Journal of the American Statistical Association, Taylor & Francis Journals, vol. 111(515), pages 1266-1277, July.
    5. Yanyuan Ma & Jeffrey D. Hart & Ryan Janicki & Raymond J. Carroll, 2011. "Local and omnibus goodness‐of‐fit tests in classical measurement error models," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 73(1), pages 81-98, January.
    6. Simon N. Vandekar & Philip T. Reiss & Russell T. Shinohara, 2019. "Interpretable High-Dimensional Inference Via Score Projection With an Application in Neuroimaging," Journal of the American Statistical Association, Taylor & Francis Journals, vol. 114(526), pages 820-830, April.
    7. Alexandre Belloni & Mathieu Rosenbaum & Alexandre B. Tsybakov, 2017. "Linear and conic programming estimators in high dimensional errors-in-variables models," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 79(3), pages 939-956, June.
    8. Pixu Shi & Yuchen Zhou & Anru R Zhang, 2022. "High-dimensional log-error-in-variable regression with applications to microbial compositional data analysis [Log contrast models for experiments with mixtures]," Biometrika, Biometrika Trust, vol. 109(2), pages 405-420.
    9. Staudenmayer, John & Buonaccorsi, John P., 2005. "Measurement Error in Linear Autoregressive Models," Journal of the American Statistical Association, American Statistical Association, vol. 100, pages 841-852, September.
    10. Zheng, Zemin & Li, Yang & Yu, Chongxiu & Li, Gaorong, 2018. "Balanced estimation for high-dimensional measurement error models," Computational Statistics & Data Analysis, Elsevier, vol. 126(C), pages 78-91.
    11. Wei, Ying & Carroll, Raymond J., 2009. "Quantile Regression With Measurement Error," Journal of the American Statistical Association, American Statistical Association, vol. 104(487), pages 1129-1143.
    12. Fan J. & Li R., 2001. "Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties," Journal of the American Statistical Association, American Statistical Association, vol. 96, pages 1348-1360, December.
    13. Wenqin Pan & Donglin Zeng & Xihong Lin, 2009. "Estimation in Semiparametric Transition Measurement Error Models for Longitudinal Data," Biometrics, The International Biometric Society, vol. 65(3), pages 728-736, September.
    14. Xianyang Zhang & Guang Cheng, 2017. "Simultaneous Inference for High-Dimensional Linear Models," Journal of the American Statistical Association, Taylor & Francis Journals, vol. 112(518), pages 757-768, April.
    15. Carroll, Raymond J. & Delaigle, Aurore & Hall, Peter, 2009. "Nonparametric Prediction in Measurement Error Models," Journal of the American Statistical Association, American Statistical Association, vol. 104(487), pages 993-1003.
    16. Wenliang Pan & Xueqin Wang & Weinan Xiao & Hongtu Zhu, 2019. "A Generic Sure Independence Screening Procedure," Journal of the American Statistical Association, Taylor & Francis Journals, vol. 114(526), pages 928-937, April.
    17. Zou, Hui, 2006. "The Adaptive Lasso and Its Oracle Properties," Journal of the American Statistical Association, American Statistical Association, vol. 101, pages 1418-1429, December.
    18. S. M. Schennach & Yingyao Hu, 2013. "Nonparametric Identification and Semiparametric Estimation of Classical Measurement Error Models Without Side Information," Journal of the American Statistical Association, Taylor & Francis Journals, vol. 108(501), pages 177-186, March.
    19. Friedman, Jerome H. & Hastie, Trevor & Tibshirani, Rob, 2010. "Regularization Paths for Generalized Linear Models via Coordinate Descent," Journal of Statistical Software, Foundation for Open Access Statistics, vol. 33(i01).
    20. Fan, Jianqing & Feng, Yang & Song, Rui, 2011. "Nonparametric Independence Screening in Sparse Ultra-High-Dimensional Additive Models," Journal of the American Statistical Association, American Statistical Association, vol. 106(494), pages 544-557.
    21. Liming Wang & Xingxiang Li & Xiaoqing Wang & Peng Lai, 2022. "Unified mean-variance feature screening for ultrahigh-dimensional regression," Computational Statistics, Springer, vol. 37(4), pages 1887-1918, September.
    22. Liang, Hua & Li, Runze, 2009. "Variable Selection for Partially Linear Models With Measurement Errors," Journal of the American Statistical Association, American Statistical Association, vol. 104(485), pages 234-248.
    23. Zhao, Shaofei & Fu, Guifang, 2022. "Distribution-free and model-free multivariate feature screening via multivariate rank distance correlation," Journal of Multivariate Analysis, Elsevier, vol. 192(C).
    24. Peter Hall & Yanyuan Ma, 2007. "Semiparametric estimators of functional measurement error models with unknown error," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 69(3), pages 429-446, June.
    25. Zhenghui Feng & Jun Zhang & Qian Chen, 2020. "Statistical inference for linear regression models with additive distortion measurement errors," Statistical Papers, Springer, vol. 61(6), pages 2483-2509, December.
    26. Gareth M. James & Peter Radchenko, 2009. "A generalized Dantzig selector with shrinkage tuning," Biometrika, Biometrika Trust, vol. 96(2), pages 323-337.
    27. Wang, Liqun & Hsiao, Cheng, 2011. "Method of moments estimation and identifiability of semiparametric nonlinear errors-in-variables models," Journal of Econometrics, Elsevier, vol. 165(1), pages 30-44.
    28. Shujie Ma & Runze Li & Chih-Ling Tsai, 2017. "Variable Screening via Quantile Partial Correlation," Journal of the American Statistical Association, Taylor & Francis Journals, vol. 112(518), pages 650-663, April.
    29. Xinyu Zhang & Yanyuan Ma & Raymond J. Carroll, 2019. "MALMEM: model averaging in linear measurement error models," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 81(4), pages 763-779, September.
    30. Jianqing Fan & Jinchi Lv, 2008. "Sure independence screening for ultrahigh dimensional feature space," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 70(5), pages 849-911, November.
    31. Hui Zou & Trevor Hastie, 2005. "Addendum: Regularization and variable selection via the elastic net," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 67(5), pages 768-768, November.
    32. Hengjian Cui & Efang Kong, 2006. "Empirical Likelihood Confidence Region for Parameters in Semi‐linear Errors‐in‐Variables Models," Scandinavian Journal of Statistics, Danish Society for Theoretical Statistics;Finnish Statistical Society;Norwegian Statistical Association;Swedish Statistical Association, vol. 33(1), pages 153-168, March.
    33. Hui Zou & Trevor Hastie, 2005. "Regularization and variable selection via the elastic net," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 67(2), pages 301-320, April.
    34. Ma, Yanyuan & Carroll, Raymond J., 2006. "Locally Efficient Estimators for Semiparametric Models With Measurement Error," Journal of the American Statistical Association, American Statistical Association, vol. 101, pages 1465-1474, December.
    35. Cun-Hui Zhang & Stephanie S. Zhang, 2014. "Confidence intervals for low dimensional parameters in high dimensional linear models," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 76(1), pages 217-242, January.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Loann David Denis Desboulets, 2018. "A Review on Variable Selection in Regression Analysis," Econometrics, MDPI, vol. 6(4), pages 1-27, November.
    2. Dai, Linlin & Chen, Kani & Sun, Zhihua & Liu, Zhenqiu & Li, Gang, 2018. "Broken adaptive ridge regression and its asymptotic properties," Journal of Multivariate Analysis, Elsevier, vol. 168(C), pages 334-351.
    3. Wang, Christina Dan & Chen, Zhao & Lian, Yimin & Chen, Min, 2022. "Asset selection based on high frequency Sharpe ratio," Journal of Econometrics, Elsevier, vol. 227(1), pages 168-188.
    4. Peter Bühlmann & Jacopo Mandozzi, 2014. "High-dimensional variable screening and bias in subsequent inference, with an empirical comparison," Computational Statistics, Springer, vol. 29(3), pages 407-430, June.
    5. Zakariya Yahya Algamal & Muhammad Hisyam Lee, 2019. "A two-stage sparse logistic regression for optimal gene selection in high-dimensional microarray data classification," Advances in Data Analysis and Classification, Springer;German Classification Society - Gesellschaft für Klassifikation (GfKl);Japanese Classification Society (JCS);Classification and Data Analysis Group of the Italian Statistical Society (CLADAG);International Federation of Classification Societies (IFCS), vol. 13(3), pages 753-771, September.
    6. Min Chen & Yimin Lian & Zhao Chen & Zhengjun Zhang, 2017. "Sure explained variability and independence screening," Journal of Nonparametric Statistics, Taylor & Francis Journals, vol. 29(4), pages 849-883, October.
    7. Tanin Sirimongkolkasem & Reza Drikvandi, 2019. "On Regularisation Methods for Analysis of High Dimensional Data," Annals of Data Science, Springer, vol. 6(4), pages 737-763, December.
    8. Christis Katsouris, 2023. "High Dimensional Time Series Regression Models: Applications to Statistical Learning Methods," Papers 2308.16192, arXiv.org.
    9. Sheng, Ying & Wang, Qihua, 2020. "Model-free feature screening for ultrahigh dimensional classification," Journal of Multivariate Analysis, Elsevier, vol. 178(C).
    10. Jing Zhang & Haibo Zhou & Yanyan Liu & Jianwen Cai, 2021. "Conditional screening for ultrahigh-dimensional survival data in case-cohort studies," Lifetime Data Analysis: An International Journal Devoted to Statistical Methods and Applications for Time-to-Event Data, Springer, vol. 27(4), pages 632-661, October.
    11. Lu Xia & Bin Nan & Yi Li, 2023. "Debiased lasso for generalized linear models with a diverging number of covariates," Biometrics, The International Biometric Society, vol. 79(1), pages 344-357, March.
    12. Xia Zheng & Yaohua Rong & Ling Liu & Weihu Cheng, 2021. "A More Accurate Estimation of Semiparametric Logistic Regression," Mathematics, MDPI, vol. 9(19), pages 1-12, September.
    13. Zhihua Sun & Yi Liu & Kani Chen & Gang Li, 2022. "Broken adaptive ridge regression for right-censored survival data," Annals of the Institute of Statistical Mathematics, Springer;The Institute of Statistical Mathematics, vol. 74(1), pages 69-91, February.
    14. Tang, Lu & Zhou, Ling & Song, Peter X.-K., 2020. "Distributed simultaneous inference in generalized linear models via confidence distribution," Journal of Multivariate Analysis, Elsevier, vol. 176(C).
    15. Xiangyu Wang & Chenlei Leng, 2016. "High dimensional ordinary least squares projection for screening variables," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 78(3), pages 589-611, June.
    16. Tutz, Gerhard & Pößnecker, Wolfgang & Uhlmann, Lorenz, 2015. "Variable selection in general multinomial logit models," Computational Statistics & Data Analysis, Elsevier, vol. 82(C), pages 207-222.
    17. Mkhadri, Abdallah & Ouhourane, Mohamed, 2013. "An extended variable inclusion and shrinkage algorithm for correlated variables," Computational Statistics & Data Analysis, Elsevier, vol. 57(1), pages 631-644.
    18. Camila Epprecht & Dominique Guegan & Álvaro Veiga & Joel Correa da Rosa, 2017. "Variable selection and forecasting via automated methods for linear models: LASSO/adaLASSO and Autometrics," Post-Print halshs-00917797, HAL.
    19. Borup, Daniel & Christensen, Bent Jesper & Mühlbach, Nicolaj Søndergaard & Nielsen, Mikkel Slot, 2023. "Targeting predictors in random forest regression," International Journal of Forecasting, Elsevier, vol. 39(2), pages 841-868.
    20. Peter Martey Addo & Dominique Guegan & Bertrand Hassani, 2018. "Credit Risk Analysis Using Machine and Deep Learning Models," Risks, MDPI, vol. 6(2), pages 1-20, April.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jmathe:v:11:y:2023:i:14:p:3202-:d:1199385. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.