GiVA: Gradient-Informed Bases for Vector-Based Adaptation

2026-04-23Computation and Language

Computation and LanguageArtificial Intelligence
AI summary

The authors focus on making large AI models easier to fine-tune with fewer new parameters. They improve a method called vector-based adaptation by introducing GiVA, which uses gradients to better start the training process. GiVA trains as fast as a popular method called LoRA but needs much fewer resources. Tests on language and image tasks show GiVA works as well or better than other methods while being more efficient.

parameter-efficient fine-tuningLoRAvector-based adaptationgradient-based initializationranknatural language understandingnatural language generationimage classification
Authors
Neeraj Gangwar, Rishabh Deshmukh, Michael Shavlovsky, Hancao Li, Vivek Mittal, Lexing Ying, Nickvash Kani
Abstract
As model sizes continue to grow, parameter-efficient fine-tuning has emerged as a powerful alternative to full fine-tuning. While LoRA is widely adopted among these methods, recent research has explored vector-based adaptation methods due to their extreme parameter efficiency. However, these methods typically require substantially higher ranks than LoRA to match its performance, leading to increased training costs. This work introduces GiVA, a gradient-based initialization strategy for vector-based adaptation. It achieves training times comparable to LoRA and maintains the extreme parameter efficiency of vector-based adaptation. We evaluate GiVA across diverse benchmarks, including natural language understanding, natural language generation, and image classification. Experiments show that our approach consistently outperforms or achieves performance competitive with existing vector-based adaptation methods and LoRA while reducing rank requirements by a factor of eight ($8\times$).