Learning Task-Agnostic Embedding of Multiple Black-Box Experts for Multi-Task Model Fusion

Nghia Hoang, Thanh Lam, Bryan Kian Hsiang Low, Patrick Jaillet
Proceedings of the 37th International Conference on Machine Learning, PMLR 119:4282-4292, 2020.

Abstract

Model fusion is an emerging study in collective learning where heterogeneous experts with private data and learning architectures need to combine their black-box knowledge for better performance. Existing literature achieves this via a local knowledge distillation scheme that transfuses the predictive patterns of each pre-trained expert onto a white-box imitator model, which can be incorporated efficiently into a global model. This scheme however does not extend to multi-task scenarios where different experts were trained to solve different tasks and only part of their distilled knowledge is relevant to a new task. To address this multi-task challenge, we develop a new fusion paradigm that represents each expert as a distribution over a spectrum of predictive prototypes, which are isolated from task-specific information encoded within the prototype distribution. The task-agnostic prototypes can then be reintegrated to generate a new model that solves a new task encoded with a different prototype distribution. The fusion and adaptation performance of the proposed framework is demonstrated empirically on several real-world benchmark datasets.

Cite this Paper


BibTeX
@InProceedings{pmlr-v119-hoang20b, title = {Learning Task-Agnostic Embedding of Multiple Black-Box Experts for Multi-Task Model Fusion}, author = {Hoang, Nghia and Lam, Thanh and Low, Bryan Kian Hsiang and Jaillet, Patrick}, booktitle = {Proceedings of the 37th International Conference on Machine Learning}, pages = {4282--4292}, year = {2020}, editor = {III, Hal Daumé and Singh, Aarti}, volume = {119}, series = {Proceedings of Machine Learning Research}, month = {13--18 Jul}, publisher = {PMLR}, pdf = {http://proceedings.mlr.press/v119/hoang20b/hoang20b.pdf}, url = {http://proceedings.mlr.press/v119/hoang20b.html}, abstract = {Model fusion is an emerging study in collective learning where heterogeneous experts with private data and learning architectures need to combine their black-box knowledge for better performance. Existing literature achieves this via a local knowledge distillation scheme that transfuses the predictive patterns of each pre-trained expert onto a white-box imitator model, which can be incorporated efficiently into a global model. This scheme however does not extend to multi-task scenarios where different experts were trained to solve different tasks and only part of their distilled knowledge is relevant to a new task. To address this multi-task challenge, we develop a new fusion paradigm that represents each expert as a distribution over a spectrum of predictive prototypes, which are isolated from task-specific information encoded within the prototype distribution. The task-agnostic prototypes can then be reintegrated to generate a new model that solves a new task encoded with a different prototype distribution. The fusion and adaptation performance of the proposed framework is demonstrated empirically on several real-world benchmark datasets.} }
Endnote
%0 Conference Paper %T Learning Task-Agnostic Embedding of Multiple Black-Box Experts for Multi-Task Model Fusion %A Nghia Hoang %A Thanh Lam %A Bryan Kian Hsiang Low %A Patrick Jaillet %B Proceedings of the 37th International Conference on Machine Learning %C Proceedings of Machine Learning Research %D 2020 %E Hal Daumé III %E Aarti Singh %F pmlr-v119-hoang20b %I PMLR %P 4282--4292 %U http://proceedings.mlr.press/v119/hoang20b.html %V 119 %X Model fusion is an emerging study in collective learning where heterogeneous experts with private data and learning architectures need to combine their black-box knowledge for better performance. Existing literature achieves this via a local knowledge distillation scheme that transfuses the predictive patterns of each pre-trained expert onto a white-box imitator model, which can be incorporated efficiently into a global model. This scheme however does not extend to multi-task scenarios where different experts were trained to solve different tasks and only part of their distilled knowledge is relevant to a new task. To address this multi-task challenge, we develop a new fusion paradigm that represents each expert as a distribution over a spectrum of predictive prototypes, which are isolated from task-specific information encoded within the prototype distribution. The task-agnostic prototypes can then be reintegrated to generate a new model that solves a new task encoded with a different prototype distribution. The fusion and adaptation performance of the proposed framework is demonstrated empirically on several real-world benchmark datasets.
APA
Hoang, N., Lam, T., Low, B.K.H. & Jaillet, P.. (2020). Learning Task-Agnostic Embedding of Multiple Black-Box Experts for Multi-Task Model Fusion. Proceedings of the 37th International Conference on Machine Learning, in Proceedings of Machine Learning Research 119:4282-4292 Available from http://proceedings.mlr.press/v119/hoang20b.html.

Related Material