Show simple item record

Implicit Regularization of Gradient Descent in Realistic Settings

dc.contributor.authorMa, Jianhao
dc.date.accessioned2025-05-12T17:35:13Z
dc.date.available2025-05-12T17:35:13Z
dc.date.issued2025
dc.date.submitted2025
dc.identifier.urihttps://hdl.handle.net/2027.42/197098
dc.description.abstractGradient descent (GD) is the backbone of modern machine learning optimization, and its success is often attributed not just to efficiency but also to an intriguing phenomenon known as implicit regularization—the tendency of GD to favor solutions that generalize well, even without explicit constraints. While existing theoretical studies have shed light on this behavior, they frequently rely on idealized assumptions, such as isotropic data distributions or benign optimization landscapes, which rarely hold in practice. This gap highlights the need to understand implicit regularization in more realistic and challenging settings that better reflect practical machine learning problems. This dissertation explores implicit regularization in gradient-based learning under three key challenges: (1) robustness to heavy-tailed outlier noise, (2) learning with non-isotropic input distributions, and (3) developing a general theoretical framework for characterizing optimization trajectories. First, we study robust learning with $ell_1$-loss, including robust matrix sensing and robust deep linear networks, and show that GD implicitly preserves low-rank structures throughout training, achieving near-linear convergence under near-optimal sample complexity. Interestingly, we prove that the ground truth corresponds to a strict saddle point, countering the conventional wisdom that saddle points inherently impede optimization. Second, we analyze learning with non-isotropic input distributions—a broad category that includes unregularized matrix completion as a special case. We introduce a novel statistical decoupling technique that establishes near-optimal sample complexity guarantees for GD without relying on standard isotropic assumptions. Lastly, we develop a unified framework based on structured basis function decomposition, revealing that GD trajectories, when projected onto a suitable function basis, exhibit an almost monotonic progression despite their apparent complexity. Altogether, this work advances the theoretical understanding of implicit regularization under realistic conditions, offering rigorous insights into how gradient descent selects generalizable solutions. These results have broad implications for designing robust and efficient learning algorithms across diverse machine learning tasks.
dc.language.isoen_US
dc.subjectGradient descent, implicit regularization, continuous optimization
dc.titleImplicit Regularization of Gradient Descent in Realistic Settings
dc.typeThesis
dc.description.thesisdegreenamePhD
dc.description.thesisdegreedisciplineIndustrial & Operations Engineering
dc.description.thesisdegreegrantorUniversity of Michigan, Horace H. Rackham School of Graduate Studies
dc.contributor.committeememberFattahi, Salar
dc.contributor.committeememberHu, Wei
dc.contributor.committeememberBerahas, Albert Solomon
dc.contributor.committeememberLee, Jon
dc.subject.hlbsecondlevelIndustrial and Operations Engineering
dc.subject.hlbtoplevelEngineering
dc.contributor.affiliationumcampusAnn Arbor
dc.description.bitstreamurlhttp://deepblue.lib.umich.edu/bitstream/2027.42/197098/1/jianhao_1.pdf
dc.identifier.doihttps://dx.doi.org/10.7302/25524
dc.identifier.orcid0000-0003-1440-6645
dc.identifier.name-orcidMa, Jianhao; 0000-0003-1440-6645en_US
dc.working.doi10.7302/25524en
dc.owningcollnameDissertations and Theses (Ph.D. and Master's)


Files in this item

Show simple item record

Remediation of Harmful Language

The University of Michigan Library aims to describe its collections in a way that respects the people and communities who create, use, and are represented in them. We encourage you to Contact Us anonymously if you encounter harmful or problematic language in catalog records or finding aids. More information about our policies and practices is available at Remediation of Harmful Language.

Accessibility

If you are unable to use this file in its current format, please select the Contact Us link and we can modify it to make it more accessible to you.