Show simple item record

Investigation of Smooth and Non-smooth Penalties for Regularized Model Selection in Regression.

dc.contributor.authorChoi, Nam Heeen_US
dc.date.accessioned2010-01-07T16:24:27Z
dc.date.availableNO_RESTRICTIONen_US
dc.date.available2010-01-07T16:24:27Z
dc.date.issued2009en_US
dc.date.submitteden_US
dc.identifier.urihttps://hdl.handle.net/2027.42/64649
dc.description.abstractIn this thesis, new approaches for using regularized regression in model selection are proposed, and we characterize the circumstances in which regularized regression improves our ability to discriminate models. First, we propose a variable selection method for regression models with interactions, using L1 regularization to automatically enforces heredity constraints. Theoretical study shows that asymptotically the proposed method performs as well as when the true model is known in advance under some regularity conditions. Numerical results show that the method performs favorably in terms of prediction and variable selection compared to some other recently developed methods. Second, regularized regression methods including ridge regression, the Lasso and the elastic net are investigated in terms of their abilities to rank the predictors in a regression model based on the sizes of their effects. Intuitively, regularization should be most useful when strong collinearity is present, however, we find that not all models with collinearity benefit from regularization. We were able to characterize situations in which regularization is either helpful, harmful, or neutral for ranking performance, and defined a sense in which regularization improves performance more often than not. By analytical and numerical studies, we show that L2-regularization outperforms L1-regularization for ranking performance, especially when the effects are weak, partly because when univariate analysis is optimal, ridge regression can better approximate univariate analysis than the Lasso. Our results also imply that the best regression estimator for variable ranking and for prediction may differ. This work may have implications for genetic mapping and other analyses involving regression methods with weak effects and collinear regressors.en_US
dc.format.extent3004823 bytes
dc.format.extent1373 bytes
dc.format.mimetypeapplication/pdf
dc.format.mimetypetext/plain
dc.language.isoen_USen_US
dc.subjectRegularizationen_US
dc.subjectPenalized Regressionen_US
dc.subjectLassoen_US
dc.subjectRidge Regressionen_US
dc.subjectHeredityen_US
dc.subjectRankingen_US
dc.titleInvestigation of Smooth and Non-smooth Penalties for Regularized Model Selection in Regression.en_US
dc.typeThesisen_US
dc.description.thesisdegreenamePhDen_US
dc.description.thesisdegreedisciplineStatisticsen_US
dc.description.thesisdegreegrantorUniversity of Michigan, Horace H. Rackham School of Graduate Studiesen_US
dc.contributor.committeememberShedden, Kerby A.en_US
dc.contributor.committeememberZhu, Jien_US
dc.contributor.committeememberStoev, Stilian Atanasoven_US
dc.contributor.committeememberTaylor, Jeremy M.en_US
dc.subject.hlbsecondlevelStatistics and Numeric Dataen_US
dc.subject.hlbtoplevelScienceen_US
dc.description.bitstreamurlhttp://deepblue.lib.umich.edu/bitstream/2027.42/64649/1/nami_1.pdf
dc.owningcollnameDissertations and Theses (Ph.D. and Master's)


Files in this item

Show simple item record

Remediation of Harmful Language

The University of Michigan Library aims to describe library materials in a way that respects the people and communities who create, use, and are represented in our collections. Report harmful or offensive language in catalog records, finding aids, or elsewhere in our collections anonymously through our metadata feedback form. More information at Remediation of Harmful Language.

Accessibility

If you are unable to use this file in its current format, please select the Contact Us link and we can modify it to make it more accessible to you.