Investigation of Smooth and Non-smooth Penalties for Regularized Model Selection in Regression.
dc.contributor.author | Choi, Nam Hee | en_US |
dc.date.accessioned | 2010-01-07T16:24:27Z | |
dc.date.available | NO_RESTRICTION | en_US |
dc.date.available | 2010-01-07T16:24:27Z | |
dc.date.issued | 2009 | en_US |
dc.date.submitted | en_US | |
dc.identifier.uri | https://hdl.handle.net/2027.42/64649 | |
dc.description.abstract | In this thesis, new approaches for using regularized regression in model selection are proposed, and we characterize the circumstances in which regularized regression improves our ability to discriminate models. First, we propose a variable selection method for regression models with interactions, using L1 regularization to automatically enforces heredity constraints. Theoretical study shows that asymptotically the proposed method performs as well as when the true model is known in advance under some regularity conditions. Numerical results show that the method performs favorably in terms of prediction and variable selection compared to some other recently developed methods. Second, regularized regression methods including ridge regression, the Lasso and the elastic net are investigated in terms of their abilities to rank the predictors in a regression model based on the sizes of their effects. Intuitively, regularization should be most useful when strong collinearity is present, however, we find that not all models with collinearity benefit from regularization. We were able to characterize situations in which regularization is either helpful, harmful, or neutral for ranking performance, and defined a sense in which regularization improves performance more often than not. By analytical and numerical studies, we show that L2-regularization outperforms L1-regularization for ranking performance, especially when the effects are weak, partly because when univariate analysis is optimal, ridge regression can better approximate univariate analysis than the Lasso. Our results also imply that the best regression estimator for variable ranking and for prediction may differ. This work may have implications for genetic mapping and other analyses involving regression methods with weak effects and collinear regressors. | en_US |
dc.format.extent | 3004823 bytes | |
dc.format.extent | 1373 bytes | |
dc.format.mimetype | application/pdf | |
dc.format.mimetype | text/plain | |
dc.language.iso | en_US | en_US |
dc.subject | Regularization | en_US |
dc.subject | Penalized Regression | en_US |
dc.subject | Lasso | en_US |
dc.subject | Ridge Regression | en_US |
dc.subject | Heredity | en_US |
dc.subject | Ranking | en_US |
dc.title | Investigation of Smooth and Non-smooth Penalties for Regularized Model Selection in Regression. | en_US |
dc.type | Thesis | en_US |
dc.description.thesisdegreename | PhD | en_US |
dc.description.thesisdegreediscipline | Statistics | en_US |
dc.description.thesisdegreegrantor | University of Michigan, Horace H. Rackham School of Graduate Studies | en_US |
dc.contributor.committeemember | Shedden, Kerby A. | en_US |
dc.contributor.committeemember | Zhu, Ji | en_US |
dc.contributor.committeemember | Stoev, Stilian Atanasov | en_US |
dc.contributor.committeemember | Taylor, Jeremy M. | en_US |
dc.subject.hlbsecondlevel | Statistics and Numeric Data | en_US |
dc.subject.hlbtoplevel | Science | en_US |
dc.description.bitstreamurl | http://deepblue.lib.umich.edu/bitstream/2027.42/64649/1/nami_1.pdf | |
dc.owningcollname | Dissertations and Theses (Ph.D. and Master's) |
Files in this item
Remediation of Harmful Language
The University of Michigan Library aims to describe library materials in a way that respects the people and communities who create, use, and are represented in our collections. Report harmful or offensive language in catalog records, finding aids, or elsewhere in our collections anonymously through our metadata feedback form. More information at Remediation of Harmful Language.
Accessibility
If you are unable to use this file in its current format, please select the Contact Us link and we can modify it to make it more accessible to you.