The Fifth Elephant 2025 Annual Conference CfP
Speak at The Fifth Elephant 2025 Annual Conference
Submitted May 30, 2025
Large pre-trained models are now the norm, making Parameter-Efficient Fine-Tuning techniques like LoRA essential to reduce computational and storage costs. But why do these methods work so well? This talk explores the theory of Intrinsic Dimension (ID)—the idea that neural networks often need far fewer effective directions to learn a task than their total parameters suggest.
We’ll estimate a task’s ID via random subspace training on an MLP for MNIST, reproducing results from foundational papers. Then, we’ll compare how LoRA approximates subspace training in compute, training time, and accuracy—clarifying key design trade-offs. LoRA succeeds not just from engineering but by exploiting the low-dimensional structure revealed by ID.
We also highlight PyTorch internals that enable flexible subspace training. This talk builds on a four-part blog series bridging theory and engineering.
This talk builds on the four-part blog series on LoRA and Intrinsic Dimension). These blogs gained good visibility — receiving positive traction on /r/MachineLearning and ranking 7th on Hacker News (front page for a day).
While the first two blogs on LoRA became the basis of the talk at PyCon India 2024 and at the Fifth Elephant Open Source AI meet (April 2025), this submission is based on the later two blogs which dive deeper into intrinsic dimension and measuring model complexity offering new perspectives.
{{ gettext('Login to leave a comment') }}
{{ gettext('Post a comment…') }}{{ errorMsg }}
{{ gettext('No comments posted yet') }}