Boosting Multi-Domain Fine-Tuning of Large Language Models through Evolving Interactions between Samples

Xize Liang, Lin Yang, Jie Wang, Yiyang Lu, Runyu Wu, Hanzhu Chen, Jianye Hao
Proceedings of the 42nd International Conference on Machine Learning, PMLR 267:37427-37441, 2025.

Abstract

The multi-domain fine-tuning of large language models (LLMs) confronts a notorious trade-off among abilities across domains. Existing studies attribute this trade-off to the conflicts between samples rooted in inherent semantics. Recent approaches attempt to mitigate these conflicts through the empirical investigation or heuristic strategies. However, without a fundamental understanding of interactions between samples, they yield only marginal improvements, while incurring substantial trial-and-error costs. To address this challenge, we move beyond empirical studies by modeling interactions between samples as their influence on each other’s loss, estimated using gradients. Intriguingly, we find that these interactions evolve throughout training rather than being purely determined by inherent semantics. Building on this insight, we propose EVolving Interaction-guided Curriculum (EVIC), which iteratively selects samples that positively influence the overall dataset for training. By dynamically adapting the training curriculum to prioritize samples that contribute the most to the model training, EVIC effectively mitigates conflicts and improves the sample efficiency. Extensive experiments on a mixed dataset covering coding, math, and general tasks with several model architectures show that EVIC significantly outperforms all baselines across diverse capabilities.

Cite this Paper


BibTeX
@InProceedings{pmlr-v267-liang25q, title = {Boosting Multi-Domain Fine-Tuning of Large Language Models through Evolving Interactions between Samples}, author = {Liang, Xize and Yang, Lin and Wang, Jie and Lu, Yiyang and Wu, Runyu and Chen, Hanzhu and Hao, Jianye}, booktitle = {Proceedings of the 42nd International Conference on Machine Learning}, pages = {37427--37441}, year = {2025}, editor = {Singh, Aarti and Fazel, Maryam and Hsu, Daniel and Lacoste-Julien, Simon and Berkenkamp, Felix and Maharaj, Tegan and Wagstaff, Kiri and Zhu, Jerry}, volume = {267}, series = {Proceedings of Machine Learning Research}, month = {13--19 Jul}, publisher = {PMLR}, pdf = {https://raw.githubusercontent.com/mlresearch/v267/main/assets/liang25q/liang25q.pdf}, url = {https://proceedings.mlr.press/v267/liang25q.html}, abstract = {The multi-domain fine-tuning of large language models (LLMs) confronts a notorious trade-off among abilities across domains. Existing studies attribute this trade-off to the conflicts between samples rooted in inherent semantics. Recent approaches attempt to mitigate these conflicts through the empirical investigation or heuristic strategies. However, without a fundamental understanding of interactions between samples, they yield only marginal improvements, while incurring substantial trial-and-error costs. To address this challenge, we move beyond empirical studies by modeling interactions between samples as their influence on each other’s loss, estimated using gradients. Intriguingly, we find that these interactions evolve throughout training rather than being purely determined by inherent semantics. Building on this insight, we propose EVolving Interaction-guided Curriculum (EVIC), which iteratively selects samples that positively influence the overall dataset for training. By dynamically adapting the training curriculum to prioritize samples that contribute the most to the model training, EVIC effectively mitigates conflicts and improves the sample efficiency. Extensive experiments on a mixed dataset covering coding, math, and general tasks with several model architectures show that EVIC significantly outperforms all baselines across diverse capabilities.} }
Endnote
%0 Conference Paper %T Boosting Multi-Domain Fine-Tuning of Large Language Models through Evolving Interactions between Samples %A Xize Liang %A Lin Yang %A Jie Wang %A Yiyang Lu %A Runyu Wu %A Hanzhu Chen %A Jianye Hao %B Proceedings of the 42nd International Conference on Machine Learning %C Proceedings of Machine Learning Research %D 2025 %E Aarti Singh %E Maryam Fazel %E Daniel Hsu %E Simon Lacoste-Julien %E Felix Berkenkamp %E Tegan Maharaj %E Kiri Wagstaff %E Jerry Zhu %F pmlr-v267-liang25q %I PMLR %P 37427--37441 %U https://proceedings.mlr.press/v267/liang25q.html %V 267 %X The multi-domain fine-tuning of large language models (LLMs) confronts a notorious trade-off among abilities across domains. Existing studies attribute this trade-off to the conflicts between samples rooted in inherent semantics. Recent approaches attempt to mitigate these conflicts through the empirical investigation or heuristic strategies. However, without a fundamental understanding of interactions between samples, they yield only marginal improvements, while incurring substantial trial-and-error costs. To address this challenge, we move beyond empirical studies by modeling interactions between samples as their influence on each other’s loss, estimated using gradients. Intriguingly, we find that these interactions evolve throughout training rather than being purely determined by inherent semantics. Building on this insight, we propose EVolving Interaction-guided Curriculum (EVIC), which iteratively selects samples that positively influence the overall dataset for training. By dynamically adapting the training curriculum to prioritize samples that contribute the most to the model training, EVIC effectively mitigates conflicts and improves the sample efficiency. Extensive experiments on a mixed dataset covering coding, math, and general tasks with several model architectures show that EVIC significantly outperforms all baselines across diverse capabilities.
APA
Liang, X., Yang, L., Wang, J., Lu, Y., Wu, R., Chen, H. & Hao, J.. (2025). Boosting Multi-Domain Fine-Tuning of Large Language Models through Evolving Interactions between Samples. Proceedings of the 42nd International Conference on Machine Learning, in Proceedings of Machine Learning Research 267:37427-37441 Available from https://proceedings.mlr.press/v267/liang25q.html.

Related Material