Skip to Content

Search: {{$root.lsaSearchQuery.q}}, Page {{$root.page}}

AIM Seminar: Efficient Low-Dimensional Compression for Deep Overparameterized Learning and Fine-Tuning

Laura Balzano (University of Michigan, Electrical Engineering and Computer Science)
Friday, April 4, 2025
3:00-4:00 PM
1084 East Hall Map
Abstract: While overparameterization in machine learning models offers great benefits in terms of optimization and generalization, it also leads to increased computational requirements as model sizes grow. In this work, we demonstrate that we can reap the benefits of overparameterization without the computational burden. First, we develop theory showing that when training the parameters of a deep linear network to fit a low-rank or wide matrix, the gradient dynamics of each weight matrix are confined to an invariant low-dimensional subspace. This is done by carefully studying the gradient update step, which is the product of several matrix variables, and noticing the way low-rank structure passes from the low-rank target through the variables sequentially. Given this invariant subspace, we can construct and train compact, highly compressed factorizations possessing the same benefits as their overparameterized counterparts. For language model fine-tuning, we introduce a method called "Deep LoRA", which improves the existing low-rank adaptation (LoRA) technique. While this technique does not arise directly from our theory, it involves only a minor modification that is surprisingly effective and of great interest for future theoretical study.

Contact: Peter Miller
Building: East Hall
Event Type: Lecture / Discussion
Tags: Mathematics
Source: Happening @ Michigan from Applied Interdisciplinary Mathematics (AIM) Seminar - Department of Mathematics, Department of Mathematics