IDEAS home Printed from https://ideas.repec.org/a/eee/csdana/v78y2014icp176-185.html
   My bibliography  Save this article

Variable and boundary selection for functional data via multiclass logistic regression modeling

Author

Listed:
  • Matsui, Hidetoshi

Abstract

Penalties with an ℓ1 norm provide solutions in which some coefficients are exactly zero and can be used for selecting variables in regression settings. When applied to the logistic regression model, they also can be used to select variables which affect classification. We focus on the form of ℓ1 penalties in logistic regression models for functional data, in particular, their use in classifying functions into three or more groups while simultaneously selecting variables or classification boundaries. We provide penalties that appropriately select the variables in functional multiclass logistic regression models. Analysis of simulation and real data show that the form of the penalty should be selected in accordance with the purpose of the analysis.

Suggested Citation

  • Matsui, Hidetoshi, 2014. "Variable and boundary selection for functional data via multiclass logistic regression modeling," Computational Statistics & Data Analysis, Elsevier, vol. 78(C), pages 176-185.
  • Handle: RePEc:eee:csdana:v:78:y:2014:i:c:p:176-185
    DOI: 10.1016/j.csda.2014.04.015
    as

    Download full text from publisher

    File URL: http://www.sciencedirect.com/science/article/pii/S0167947314001224
    Download Restriction: Full text for ScienceDirect subscribers only.

    File URL: https://libkey.io/10.1016/j.csda.2014.04.015?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    References listed on IDEAS

    as
    1. Sadanori Konishi, 2004. "Bayesian information criteria and smoothing parameter selection in radial basis function networks," Biometrika, Biometrika Trust, vol. 91(1), pages 27-43, March.
    2. Fan J. & Li R., 2001. "Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties," Journal of the American Statistical Association, American Statistical Association, vol. 96, pages 1348-1360, December.
    3. Escabias, M. & Aguilera, A.M. & Valderrama, M.J., 2007. "Functional PLS logit regression model," Computational Statistics & Data Analysis, Elsevier, vol. 51(10), pages 4891-4902, June.
    4. Philip T. Reiss & R. Todd Ogden, 2010. "Functional Generalized Linear Models with Images as Predictors," Biometrics, The International Biometric Society, vol. 66(1), pages 61-69, March.
    5. Cardot, Hervé & Sarda, Pacal, 2005. "Estimation in generalized linear models for functional data via penalized likelihood," Journal of Multivariate Analysis, Elsevier, vol. 92(1), pages 24-41, January.
    6. F. Ferraty & P. Hall & P. Vieu, 2010. "Most-predictive design points for functional data predictors," Biometrika, Biometrika Trust, vol. 97(4), pages 807-824.
    7. Ming Yuan & Ali Ekici & Zhaosong Lu & Renato Monteiro, 2007. "Dimension reduction and coefficient estimation in multivariate linear regression," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 69(3), pages 329-346, June.
    8. Yuko Araki & Sadanori Konishi & Shuichi Kawano & Hidetoshi Matsui, 2009. "Functional regression modeling via regularized Gaussian basis expansions," Annals of the Institute of Statistical Mathematics, Springer;The Institute of Statistical Mathematics, vol. 61(4), pages 811-833, December.
    9. M. Aguilera-Morillo & Ana Aguilera & Manuel Escabias & Mariano Valderrama, 2013. "Penalized spline approaches for functional logit regression," TEST: An Official Journal of the Spanish Society of Statistics and Operations Research, Springer;Sociedad de Estadística e Investigación Operativa, vol. 22(2), pages 251-277, June.
    10. Hansheng Wang & Runze Li & Chih-Ling Tsai, 2007. "Tuning parameter selectors for the smoothly clipped absolute deviation method," Biometrika, Biometrika Trust, vol. 94(3), pages 553-568.
    11. Zou, Hui, 2006. "The Adaptive Lasso and Its Oracle Properties," Journal of the American Statistical Association, American Statistical Association, vol. 101, pages 1418-1429, December.
    12. Li, Yehua & Wang, Naisyin & Carroll, Raymond J., 2010. "Generalized Functional Linear Models With Semiparametric Single-Index Interactions," Journal of the American Statistical Association, American Statistical Association, vol. 105(490), pages 621-633.
    13. Lukas Meier & Sara Van De Geer & Peter Bühlmann, 2008. "The group lasso for logistic regression," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 70(1), pages 53-71, February.
    14. Friedman, Jerome H. & Hastie, Trevor & Tibshirani, Rob, 2010. "Regularization Paths for Generalized Linear Models via Coordinate Descent," Journal of Statistical Software, Foundation for Open Access Statistics, vol. 33(i01).
    15. Gareth M. James, 2002. "Generalized linear models with functional predictors," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 64(3), pages 411-432, August.
    16. Mingotti, Nicola & Lillo Rodríguez, Rosa Elvira & Romo, Juan, 2013. "Lasso variable selection in functional regression," DES - Working Papers. Statistics and Econometrics. WS ws131413, Universidad Carlos III de Madrid. Departamento de Estadística.
    17. Matsui, Hidetoshi & Konishi, Sadanori, 2011. "Variable selection for functional regression models via the L1 regularization," Computational Statistics & Data Analysis, Elsevier, vol. 55(12), pages 3304-3310, December.
    18. Hui Zou & Trevor Hastie, 2005. "Addendum: Regularization and variable selection via the elastic net," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 67(5), pages 768-768, November.
    19. Hui Zou & Trevor Hastie, 2005. "Regularization and variable selection via the elastic net," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 67(2), pages 301-320, April.
    20. Ming Yuan & Yi Lin, 2006. "Model selection and estimation in regression with grouped variables," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 68(1), pages 49-67, February.
    Full references (including those not matched with items on IDEAS)

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Casado Yusta, Silvia & Nœ–ez Letamendía, Laura & Pacheco Bonrostro, Joaqu’n Antonio, 2018. "Predicting Corporate Failure: The GRASP-LOGIT Model || Predicci—n de la quiebra empresarial: el modelo GRASP-LOGIT," Revista de Métodos Cuantitativos para la Economía y la Empresa = Journal of Quantitative Methods for Economics and Business Administration, Universidad Pablo de Olavide, Department of Quantitative Methods for Economics and Business Administration, vol. 26(1), pages 294-314, Diciembre.
    2. Łukasz Smaga & Hidetoshi Matsui, 2018. "A note on variable selection in functional regression via random subspace method," Statistical Methods & Applications, Springer;Società Italiana di Statistica, vol. 27(3), pages 455-477, August.
    3. Gregorutti, Baptiste & Michel, Bertrand & Saint-Pierre, Philippe, 2015. "Grouped variable importance with random forests and application to multiple functional data analysis," Computational Statistics & Data Analysis, Elsevier, vol. 90(C), pages 15-35.
    4. Fraiman, Ricardo & Gimenez, Yanina & Svarc, Marcela, 2016. "Feature selection for functional data," Journal of Multivariate Analysis, Elsevier, vol. 146(C), pages 191-208.
    5. Andreas Groll & Trevor Hastie & Gerhard Tutz, 2017. "Selection of effects in Cox frailty models by regularization methods," Biometrics, The International Biometric Society, vol. 73(3), pages 846-856, September.
    6. Floriello, Davide & Vitelli, Valeria, 2017. "Sparse clustering of functional data," Journal of Multivariate Analysis, Elsevier, vol. 154(C), pages 1-18.
    7. Mirosław Krzyśko & Łukasz Smaga, 2017. "An Application Of Functional Multivariate Regression Model To Multiclass Classification," Statistics in Transition New Series, Polish Statistical Association, vol. 18(3), pages 433-442, September.
    8. Krzyśko Mirosław & Smaga Łukasz, 2017. "An Application of Functional Multivariate Regression Model to Multiclass Classification," Statistics in Transition New Series, Polish Statistical Association, vol. 18(3), pages 433-442, September.
    9. Aneiros, Germán & Novo, Silvia & Vieu, Philippe, 2022. "Variable selection in functional regression models: A review," Journal of Multivariate Analysis, Elsevier, vol. 188(C).

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Aneiros, Germán & Novo, Silvia & Vieu, Philippe, 2022. "Variable selection in functional regression models: A review," Journal of Multivariate Analysis, Elsevier, vol. 188(C).
    2. Tutz, Gerhard & Pößnecker, Wolfgang & Uhlmann, Lorenz, 2015. "Variable selection in general multinomial logit models," Computational Statistics & Data Analysis, Elsevier, vol. 82(C), pages 207-222.
    3. Zanhua Yin, 2020. "Variable selection for sparse logistic regression," Metrika: International Journal for Theoretical and Applied Statistics, Springer, vol. 83(7), pages 821-836, October.
    4. Pei Wang & Shunjie Chen & Sijia Yang, 2022. "Recent Advances on Penalized Regression Models for Biological Data," Mathematics, MDPI, vol. 10(19), pages 1-24, October.
    5. Hirose, Kei & Tateishi, Shohei & Konishi, Sadanori, 2013. "Tuning parameter selection in sparse regression modeling," Computational Statistics & Data Analysis, Elsevier, vol. 59(C), pages 28-40.
    6. Capanu, Marinela & Giurcanu, Mihai & Begg, Colin B. & Gönen, Mithat, 2023. "Subsampling based variable selection for generalized linear models," Computational Statistics & Data Analysis, Elsevier, vol. 184(C).
    7. Loann David Denis Desboulets, 2018. "A Review on Variable Selection in Regression Analysis," Econometrics, MDPI, vol. 6(4), pages 1-27, November.
    8. Qingliang Fan & Yaqian Wu, 2020. "Endogenous Treatment Effect Estimation with some Invalid and Irrelevant Instruments," Papers 2006.14998, arXiv.org.
    9. Yanfang Zhang & Chuanhua Wei & Xiaolin Liu, 2022. "Group Logistic Regression Models with l p,q Regularization," Mathematics, MDPI, vol. 10(13), pages 1-15, June.
    10. Satre-Meloy, Aven, 2019. "Investigating structural and occupant drivers of annual residential electricity consumption using regularization in regression models," Energy, Elsevier, vol. 174(C), pages 148-168.
    11. Young Joo Yoon & Cheolwoo Park & Erik Hofmeister & Sangwook Kang, 2012. "Group variable selection in cardiopulmonary cerebral resuscitation data for veterinary patients," Journal of Applied Statistics, Taylor & Francis Journals, vol. 39(7), pages 1605-1621, January.
    12. Haibin Zhang & Juan Wei & Meixia Li & Jie Zhou & Miantao Chao, 2014. "On proximal gradient method for the convex problems regularized with the group reproducing kernel norm," Journal of Global Optimization, Springer, vol. 58(1), pages 169-188, January.
    13. Zhang, Yan-Qing & Tian, Guo-Liang & Tang, Nian-Sheng, 2016. "Latent variable selection in structural equation models," Journal of Multivariate Analysis, Elsevier, vol. 152(C), pages 190-205.
    14. Tanin Sirimongkolkasem & Reza Drikvandi, 2019. "On Regularisation Methods for Analysis of High Dimensional Data," Annals of Data Science, Springer, vol. 6(4), pages 737-763, December.
    15. Fang, Xiaolei & Paynabar, Kamran & Gebraeel, Nagi, 2017. "Multistream sensor fusion-based prognostics model for systems with single failure modes," Reliability Engineering and System Safety, Elsevier, vol. 159(C), pages 322-331.
    16. Daniel, Jeffrey & Horrocks, Julie & Umphrey, Gary J., 2018. "Penalized composite likelihoods for inhomogeneous Gibbs point process models," Computational Statistics & Data Analysis, Elsevier, vol. 124(C), pages 104-116.
    17. van Erp, Sara & Oberski, Daniel L. & Mulder, Joris, 2018. "Shrinkage priors for Bayesian penalized regression," OSF Preprints cg8fq, Center for Open Science.
    18. Belli, Edoardo, 2022. "Smoothly adaptively centered ridge estimator," Journal of Multivariate Analysis, Elsevier, vol. 189(C).
    19. Diego Vidaurre & Concha Bielza & Pedro Larrañaga, 2013. "A Survey of L1 Regression," International Statistical Review, International Statistical Institute, vol. 81(3), pages 361-387, December.
    20. Matsui, Hidetoshi & Konishi, Sadanori, 2011. "Variable selection for functional regression models via the L1 regularization," Computational Statistics & Data Analysis, Elsevier, vol. 55(12), pages 3304-3310, December.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:eee:csdana:v:78:y:2014:i:c:p:176-185. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Catherine Liu (email available below). General contact details of provider: http://www.elsevier.com/locate/csda .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.