Exploring Minimally Sufficient Representation in Active Learning through Label-Irrelevant Patch Augmentation

Zhiyu Xue, Yinlong Dai, Qi Lei
Conference on Parsimony and Learning, PMLR 234:419-439, 2024.

Abstract

Deep learning models, which require abundant labeled data for training, are expensive and time-consuming to implement, particularly in medical imaging. Active learning (AL) aims to maximize model performance with few labeled samples by gradually expanding and labeling a new training set. In this work, we intend to learn a "good" feature representation that is both sufficient and minimal, facilitating effective AL for medical image classification. This work proposes an efficient AL framework based on off-the-shelf self-supervised learning models, complemented by a label-irrelevant patch augmentation scheme. This scheme is designed to reduce redundancy in the learned features and mitigate overfitting in the progress of AL. Our framework offers efficiency to AL in terms of parameters, samples, and computational costs. The benefits of this approach are extensively validated across various medical image classification tasks employing different AL strategies. \footnote{Source Codes: \url{https://github.com/chrisyxue/DA4AL}}.

Cite this Paper


BibTeX
@InProceedings{pmlr-v234-xue24a, title = {Exploring Minimally Sufficient Representation in Active Learning through Label-Irrelevant Patch Augmentation}, author = {Xue, Zhiyu and Dai, Yinlong and Lei, Qi}, booktitle = {Conference on Parsimony and Learning}, pages = {419--439}, year = {2024}, editor = {Chi, Yuejie and Dziugaite, Gintare Karolina and Qu, Qing and Wang, Atlas Wang and Zhu, Zhihui}, volume = {234}, series = {Proceedings of Machine Learning Research}, month = {03--06 Jan}, publisher = {PMLR}, pdf = {https://proceedings.mlr.press/v234/xue24a/xue24a.pdf}, url = {https://proceedings.mlr.press/v234/xue24a.html}, abstract = {Deep learning models, which require abundant labeled data for training, are expensive and time-consuming to implement, particularly in medical imaging. Active learning (AL) aims to maximize model performance with few labeled samples by gradually expanding and labeling a new training set. In this work, we intend to learn a "good" feature representation that is both sufficient and minimal, facilitating effective AL for medical image classification. This work proposes an efficient AL framework based on off-the-shelf self-supervised learning models, complemented by a label-irrelevant patch augmentation scheme. This scheme is designed to reduce redundancy in the learned features and mitigate overfitting in the progress of AL. Our framework offers efficiency to AL in terms of parameters, samples, and computational costs. The benefits of this approach are extensively validated across various medical image classification tasks employing different AL strategies. \footnote{Source Codes: \url{https://github.com/chrisyxue/DA4AL}}.} }
Endnote
%0 Conference Paper %T Exploring Minimally Sufficient Representation in Active Learning through Label-Irrelevant Patch Augmentation %A Zhiyu Xue %A Yinlong Dai %A Qi Lei %B Conference on Parsimony and Learning %C Proceedings of Machine Learning Research %D 2024 %E Yuejie Chi %E Gintare Karolina Dziugaite %E Qing Qu %E Atlas Wang Wang %E Zhihui Zhu %F pmlr-v234-xue24a %I PMLR %P 419--439 %U https://proceedings.mlr.press/v234/xue24a.html %V 234 %X Deep learning models, which require abundant labeled data for training, are expensive and time-consuming to implement, particularly in medical imaging. Active learning (AL) aims to maximize model performance with few labeled samples by gradually expanding and labeling a new training set. In this work, we intend to learn a "good" feature representation that is both sufficient and minimal, facilitating effective AL for medical image classification. This work proposes an efficient AL framework based on off-the-shelf self-supervised learning models, complemented by a label-irrelevant patch augmentation scheme. This scheme is designed to reduce redundancy in the learned features and mitigate overfitting in the progress of AL. Our framework offers efficiency to AL in terms of parameters, samples, and computational costs. The benefits of this approach are extensively validated across various medical image classification tasks employing different AL strategies. \footnote{Source Codes: \url{https://github.com/chrisyxue/DA4AL}}.
APA
Xue, Z., Dai, Y. & Lei, Q.. (2024). Exploring Minimally Sufficient Representation in Active Learning through Label-Irrelevant Patch Augmentation. Conference on Parsimony and Learning, in Proceedings of Machine Learning Research 234:419-439 Available from https://proceedings.mlr.press/v234/xue24a.html.

Related Material