ICon: In-Context Contribution for Automatic Data Selection
Yixin Yang, Qingxiu Dong, Linli Yao, Fangwei Zhu, Zhifang Sui
2025-05-09
Summary
This paper talks about ICon, a new way to pick out the most helpful pieces of data to use when teaching large language models, making them learn more efficiently and effectively.
What's the problem?
The problem is that not all training data is equally useful when improving AI models. Some examples help the model learn a lot, while others don't make much difference. Traditional ways to find the best data can be slow or not very accurate.
What's the solution?
The researchers introduced ICon, which is a method that figures out which pieces of data make the biggest impact on the model's learning, but does this without needing to calculate gradients, which saves time and computer power. It turns out this method works better than older techniques that rely on gradients or simple rules.
Why it matters?
This matters because it helps AI models get smarter faster and with less effort, which can lead to better performance in all sorts of applications, from chatbots to search engines. It also makes training large models more practical and cost-effective.
Abstract
In-context Learning for Contribution Measurement (ICon) is a gradient-free method that identifies high-contribution data for instruction tuning of Large Language Models, outperforming gradient-based and heuristic-based selection methods.