[edit]
Boosting Multi-Domain Fine-Tuning of Large Language Models through Evolving Interactions between Samples
Proceedings of the 42nd International Conference on Machine Learning, PMLR 267:37427-37441, 2025.
Abstract
The multi-domain fine-tuning of large language models (LLMs) confronts a notorious trade-off among abilities across domains. Existing studies attribute this trade-off to the conflicts between samples rooted in inherent semantics. Recent approaches attempt to mitigate these conflicts through the empirical investigation or heuristic strategies. However, without a fundamental understanding of interactions between samples, they yield only marginal improvements, while incurring substantial trial-and-error costs. To address this challenge, we move beyond empirical studies by modeling interactions between samples as their influence on each other’s loss, estimated using gradients. Intriguingly, we find that these interactions evolve throughout training rather than being purely determined by inherent semantics. Building on this insight, we propose EVolving Interaction-guided Curriculum (EVIC), which iteratively selects samples that positively influence the overall dataset for training. By dynamically adapting the training curriculum to prioritize samples that contribute the most to the model training, EVIC effectively mitigates conflicts and improves the sample efficiency. Extensive experiments on a mixed dataset covering coding, math, and general tasks with several model architectures show that EVIC significantly outperforms all baselines across diverse capabilities.