Reinforcement Learning Finetunes Small Subnetworks in Large Language Models
Sagnik Mukherjee, Lifan Yuan, Dilek Hakkani-Tur, Hao Peng
2025-05-23
Summary
This paper talks about a way to make big language models better by only changing a small part of them using reinforcement learning.
What's the problem?
The problem is that updating or improving large language models usually takes a lot of computer power and time because you have to adjust so many parts of the model.
What's the solution?
The researchers found that they could use reinforcement learning to fine-tune just a small section, or subnetwork, of the whole model, instead of the entire thing. This works without needing any special tricks to make the changes sparse or limited.
Why it matters?
This is important because it means we can make these huge models smarter and more useful much more efficiently, saving both time and resources.
Abstract
Reinforcement learning improves large language models with minimal parameter updates, affecting only a small subnetwork without explicit sparsity techniques.