IDEAS home Printed from https://ideas.repec.org/a/eee/csdana/v63y2013icp42-49.html
   My bibliography  Save this article

Robust variable selection through MAVE

Author

Listed:
  • Yao, Weixin
  • Wang, Qin

Abstract

Dimension reduction and variable selection play important roles in high dimensional data analysis. The sparse MAVE, a model-free variable selection method, is a nice combination of shrinkage estimation, Lasso, and an effective dimension reduction method, MAVE (minimum average variance estimation). However, it is not robust to outliers in the dependent variable because of the use of least-squares criterion. A robust variable selection method based on sparse MAVE is developed, together with an efficient estimation algorithm to enhance its practical applicability. In addition, a robust cross-validation is also proposed to select the structural dimension. The effectiveness of the new approach is verified through simulation studies and a real data analysis.

Suggested Citation

  • Yao, Weixin & Wang, Qin, 2013. "Robust variable selection through MAVE," Computational Statistics & Data Analysis, Elsevier, vol. 63(C), pages 42-49.
  • Handle: RePEc:eee:csdana:v:63:y:2013:i:c:p:42-49
    DOI: 10.1016/j.csda.2013.01.021
    as

    Download full text from publisher

    File URL: http://www.sciencedirect.com/science/article/pii/S0167947313000364
    Download Restriction: Full text for ScienceDirect subscribers only.

    File URL: https://libkey.io/10.1016/j.csda.2013.01.021?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    References listed on IDEAS

    as
    1. Cizek, P. & Hardle, W., 2006. "Robust estimation of dimension reduction space," Computational Statistics & Data Analysis, Elsevier, vol. 51(2), pages 545-555, November.
    2. Lexin Li, 2007. "Sparse sufficient dimension reduction," Biometrika, Biometrika Trust, vol. 94(3), pages 603-613.
    3. Zhu, Yu & Zeng, Peng, 2006. "Fourier Methods for Estimating the Central Subspace and the Central Mean Subspace in Regression," Journal of the American Statistical Association, American Statistical Association, vol. 101, pages 1638-1651, December.
    4. Liqiang Ni & R. Dennis Cook & Chih-Ling Tsai, 2005. "A note on shrinkage sliced inverse regression," Biometrika, Biometrika Trust, vol. 92(1), pages 242-247, March.
    5. Friedman, Jerome H. & Hastie, Trevor & Tibshirani, Rob, 2010. "Regularization Paths for Generalized Linear Models via Coordinate Descent," Journal of Statistical Software, Foundation for Open Access Statistics, vol. 33(i01).
    6. Yingcun Xia & Howell Tong & W. K. Li & Li‐Xing Zhu, 2002. "An adaptive estimation of dimension reduction space," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 64(3), pages 363-410, August.
    7. Howard D. Bondell & Lexin Li, 2009. "Shrinkage inverse regression estimation for model‐free variable selection," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 71(1), pages 287-299, January.
    8. Fan J. & Li R., 2001. "Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties," Journal of the American Statistical Association, American Statistical Association, vol. 96, pages 1348-1360, December.
    9. Ye Z. & Weiss R.E., 2003. "Using the Bootstrap to Select One of a New Class of Dimension Reduction Methods," Journal of the American Statistical Association, American Statistical Association, vol. 98, pages 968-979, January.
    10. Lexin Li & R. Dennis Cook & Christopher J. Nachtsheim, 2005. "Model‐free variable selection," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 67(2), pages 285-299, April.
    11. Wang, Qin & Yin, Xiangrong, 2008. "A nonlinear multi-dimensional variable selection method for high dimensional data: Sparse MAVE," Computational Statistics & Data Analysis, Elsevier, vol. 52(9), pages 4512-4520, May.
    12. Hui Zou & Trevor Hastie, 2005. "Addendum: Regularization and variable selection via the elastic net," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 67(5), pages 768-768, November.
    13. Hui Zou & Trevor Hastie, 2005. "Regularization and variable selection via the elastic net," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 67(2), pages 301-320, April.
    Full references (including those not matched with items on IDEAS)

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Zhang, Jing & Wang, Qin & Mays, D'Arcy, 2021. "Robust MAVE through nonconvex penalized regression," Computational Statistics & Data Analysis, Elsevier, vol. 160(C).
    2. Dernoncourt, David & Hanczar, Blaise & Zucker, Jean-Daniel, 2014. "Analysis of feature selection stability on high dimension and small sample data," Computational Statistics & Data Analysis, Elsevier, vol. 71(C), pages 681-693.
    3. Lv, Jing & Yang, Hu & Guo, Chaohui, 2015. "An efficient and robust variable selection method for longitudinal generalized linear models," Computational Statistics & Data Analysis, Elsevier, vol. 82(C), pages 74-88.
    4. Rekabdarkolaee, Hossein Moradi & Boone, Edward & Wang, Qin, 2017. "Robust estimation and variable selection in sufficient dimension reduction," Computational Statistics & Data Analysis, Elsevier, vol. 108(C), pages 146-157.

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Moradi Rekabdarkolaee, Hossein & Wang, Qin, 2017. "Variable selection through adaptive MAVE," Statistics & Probability Letters, Elsevier, vol. 128(C), pages 44-51.
    2. Wang, Tao & Zhu, Lixing, 2013. "Sparse sufficient dimension reduction using optimal scoring," Computational Statistics & Data Analysis, Elsevier, vol. 57(1), pages 223-232.
    3. Weng, Jiaying, 2022. "Fourier transform sparse inverse regression estimators for sufficient variable selection," Computational Statistics & Data Analysis, Elsevier, vol. 168(C).
    4. Wang, Qin & Yin, Xiangrong, 2008. "A nonlinear multi-dimensional variable selection method for high dimensional data: Sparse MAVE," Computational Statistics & Data Analysis, Elsevier, vol. 52(9), pages 4512-4520, May.
    5. Changrong Yan & Dixin Zhang, 2013. "Sparse dimension reduction for survival data," Computational Statistics, Springer, vol. 28(4), pages 1835-1852, August.
    6. Bilin Zeng & Xuerong Meggie Wen & Lixing Zhu, 2017. "A link-free sparse group variable selection method for single-index model," Journal of Applied Statistics, Taylor & Francis Journals, vol. 44(13), pages 2388-2400, October.
    7. Wang, Pei & Yin, Xiangrong & Yuan, Qingcong & Kryscio, Richard, 2021. "Feature filter for estimating central mean subspace and its sparse solution," Computational Statistics & Data Analysis, Elsevier, vol. 163(C).
    8. Yunquan Song & Zitong Li & Minglu Fang, 2022. "Robust Variable Selection Based on Penalized Composite Quantile Regression for High-Dimensional Single-Index Models," Mathematics, MDPI, vol. 10(12), pages 1-17, June.
    9. Wang, Qin & Yin, Xiangrong, 2008. "Sufficient dimension reduction and variable selection for regression mean function with two types of predictors," Statistics & Probability Letters, Elsevier, vol. 78(16), pages 2798-2803, November.
    10. Wei Sun & Lexin Li, 2012. "Multiple Loci Mapping via Model-free Variable Selection," Biometrics, The International Biometric Society, vol. 68(1), pages 12-22, March.
    11. Heng-Hui Lue & Bing-Ran You, 2013. "High-dimensional regression analysis with treatment comparisons," Computational Statistics, Springer, vol. 28(3), pages 1299-1317, June.
    12. Tutz, Gerhard & Pößnecker, Wolfgang & Uhlmann, Lorenz, 2015. "Variable selection in general multinomial logit models," Computational Statistics & Data Analysis, Elsevier, vol. 82(C), pages 207-222.
    13. Camila Epprecht & Dominique Guegan & Álvaro Veiga & Joel Correa da Rosa, 2017. "Variable selection and forecasting via automated methods for linear models: LASSO/adaLASSO and Autometrics," Post-Print halshs-00917797, HAL.
    14. Zichen Zhang & Ye Eun Bae & Jonathan R. Bradley & Lang Wu & Chong Wu, 2022. "SUMMIT: An integrative approach for better transcriptomic data imputation improves causal gene identification," Nature Communications, Nature, vol. 13(1), pages 1-12, December.
    15. Peter Bühlmann & Jacopo Mandozzi, 2014. "High-dimensional variable screening and bias in subsequent inference, with an empirical comparison," Computational Statistics, Springer, vol. 29(3), pages 407-430, June.
    16. Peter Martey Addo & Dominique Guegan & Bertrand Hassani, 2018. "Credit Risk Analysis Using Machine and Deep Learning Models," Risks, MDPI, vol. 6(2), pages 1-20, April.
    17. Capanu, Marinela & Giurcanu, Mihai & Begg, Colin B. & Gönen, Mithat, 2023. "Subsampling based variable selection for generalized linear models," Computational Statistics & Data Analysis, Elsevier, vol. 184(C).
    18. Loann David Denis Desboulets, 2018. "A Review on Variable Selection in Regression Analysis," Econometrics, MDPI, vol. 6(4), pages 1-27, November.
    19. Zeyu Bian & Erica E. M. Moodie & Susan M. Shortreed & Sahir Bhatnagar, 2023. "Variable selection in regression‐based estimation of dynamic treatment regimes," Biometrics, The International Biometric Society, vol. 79(2), pages 988-999, June.
    20. Jingxuan Luo & Lili Yue & Gaorong Li, 2023. "Overview of High-Dimensional Measurement Error Regression Models," Mathematics, MDPI, vol. 11(14), pages 1-22, July.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:eee:csdana:v:63:y:2013:i:c:p:42-49. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Catherine Liu (email available below). General contact details of provider: http://www.elsevier.com/locate/csda .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.