Adaptive Ensemble Prediction for Deep Neural Networks based on Confidence Level
[edit]
Proceedings of Machine Learning Research, PMLR 89:12841293, 2019.
Abstract
Ensembling multiple predictions is a widely used technique for improving the accuracy of various machine learning tasks. One obvious drawback of ensembling is its higher execution cost during inference. In this paper, we first describe our insights on the relationship between the probability of prediction and the effect of ensembling with current deep neural networks; ensembling does not help mispredictions for inputs predicted with a high probability even when there is a nonnegligible number of mispredicted inputs. This finding motivated us to develop a way to adaptively control the ensembling. If the prediction for an input reaches a high enough probability, i.e., the output from the softmax function, on the basis of the confidence level, we stop ensembling for this input to avoid wasting computation power. We evaluated the adaptive ensembling by using various datasets and showed that it reduces the computation cost significantly while achieving accuracy similar to that of static ensembling using a predefined number of local predictions. We also show that our statistically rigorous confidencelevelbased earlyexit condition reduces the burden of taskdependent threshold tuning better compared with naive early exit based on a predefined threshold in addition to yielding a better accuracy with the same cost.
Related Material


