Bayesian Active Learning by Soft Mean Objective Cost of Uncertainty

Guang Zhao, Edward Dougherty, Byung-Jun Yoon, Francis J. Alexander, Xiaoning Qian
Proceedings of The 24th International Conference on Artificial Intelligence and Statistics, PMLR 130:3970-3978, 2021.

Abstract

To achieve label efficiency for training supervised learning models, pool-based active learning sequentially selects samples from a set of candidates as queries to label by optimizing an acquisition function. One category of existing methods adopts one-step-look-ahead strategies based on acquisition functions tailored with the learning objectives, for example based on the expected loss reduction (ELR) or the mean objective cost of uncertainty (MOCU) proposed recently. These active learning methods are optimal with the maximum classification error reduction when one considers a single query. However, it is well-known that there is no performance guarantee in the long run for these myopic methods. In this paper, we show that these methods are not guaranteed to converge to the optimal classifier of the true model because MOCU is not strictly concave. Moreover, we suggest a strictly concave approximation of MOCU—Soft MOCU—that can be used to define an acquisition function to guide Bayesian active learning with theoretical convergence guarantee. For training Bayesian classifiers with both synthetic and real-world data, our experiments demonstrate the superior performance of active learning by Soft MOCU compared to other existing methods.

Cite this Paper


BibTeX
@InProceedings{pmlr-v130-zhao21c, title = { Bayesian Active Learning by Soft Mean Objective Cost of Uncertainty }, author = {Zhao, Guang and Dougherty, Edward and Yoon, Byung-Jun and J. Alexander, Francis and Qian, Xiaoning}, booktitle = {Proceedings of The 24th International Conference on Artificial Intelligence and Statistics}, pages = {3970--3978}, year = {2021}, editor = {Banerjee, Arindam and Fukumizu, Kenji}, volume = {130}, series = {Proceedings of Machine Learning Research}, month = {13--15 Apr}, publisher = {PMLR}, pdf = {http://proceedings.mlr.press/v130/zhao21c/zhao21c.pdf}, url = {https://proceedings.mlr.press/v130/zhao21c.html}, abstract = { To achieve label efficiency for training supervised learning models, pool-based active learning sequentially selects samples from a set of candidates as queries to label by optimizing an acquisition function. One category of existing methods adopts one-step-look-ahead strategies based on acquisition functions tailored with the learning objectives, for example based on the expected loss reduction (ELR) or the mean objective cost of uncertainty (MOCU) proposed recently. These active learning methods are optimal with the maximum classification error reduction when one considers a single query. However, it is well-known that there is no performance guarantee in the long run for these myopic methods. In this paper, we show that these methods are not guaranteed to converge to the optimal classifier of the true model because MOCU is not strictly concave. Moreover, we suggest a strictly concave approximation of MOCU—Soft MOCU—that can be used to define an acquisition function to guide Bayesian active learning with theoretical convergence guarantee. For training Bayesian classifiers with both synthetic and real-world data, our experiments demonstrate the superior performance of active learning by Soft MOCU compared to other existing methods. } }
Endnote
%0 Conference Paper %T Bayesian Active Learning by Soft Mean Objective Cost of Uncertainty %A Guang Zhao %A Edward Dougherty %A Byung-Jun Yoon %A Francis J. Alexander %A Xiaoning Qian %B Proceedings of The 24th International Conference on Artificial Intelligence and Statistics %C Proceedings of Machine Learning Research %D 2021 %E Arindam Banerjee %E Kenji Fukumizu %F pmlr-v130-zhao21c %I PMLR %P 3970--3978 %U https://proceedings.mlr.press/v130/zhao21c.html %V 130 %X To achieve label efficiency for training supervised learning models, pool-based active learning sequentially selects samples from a set of candidates as queries to label by optimizing an acquisition function. One category of existing methods adopts one-step-look-ahead strategies based on acquisition functions tailored with the learning objectives, for example based on the expected loss reduction (ELR) or the mean objective cost of uncertainty (MOCU) proposed recently. These active learning methods are optimal with the maximum classification error reduction when one considers a single query. However, it is well-known that there is no performance guarantee in the long run for these myopic methods. In this paper, we show that these methods are not guaranteed to converge to the optimal classifier of the true model because MOCU is not strictly concave. Moreover, we suggest a strictly concave approximation of MOCU—Soft MOCU—that can be used to define an acquisition function to guide Bayesian active learning with theoretical convergence guarantee. For training Bayesian classifiers with both synthetic and real-world data, our experiments demonstrate the superior performance of active learning by Soft MOCU compared to other existing methods.
APA
Zhao, G., Dougherty, E., Yoon, B., J. Alexander, F. & Qian, X.. (2021). Bayesian Active Learning by Soft Mean Objective Cost of Uncertainty . Proceedings of The 24th International Conference on Artificial Intelligence and Statistics, in Proceedings of Machine Learning Research 130:3970-3978 Available from https://proceedings.mlr.press/v130/zhao21c.html.

Related Material