Learning Curves of Classification Metrics based on Confusion Matrices

Yan Xue, Ruibo Wang, Xuefei Cao, Jing Yang, Jihong Li
Proceedings of the 17th Asian Conference on Machine Learning, PMLR 304:1246-1261, 2025.

Abstract

Learning curves of classification metrics, including test error, precision (P), recall (R), F$_1$ score, with regard to training set sizes are a recent hot topic in developing an advanced methodology of model selection and hyperparameter optimization. The existing studies concentrated on formulating the functional shapes of the well-behaved learning curves of test error by using a normality assumption. However, the normality assumption is unreasonable for learning curves of classification metrics because the distributions of most classification metrics, such as P, R, and F$_1$ score, are skewed, and interval estimations of the metrics based on the normality assumption may exceed [0,1]. In this study, considering most classification metrics are obtained from confusion matrices, we develop a novel method to formulate the learning curves of classification metrics by considering that the four entries in a confusion matrix jointly follow a multi-nomial distribution rather than a normality distribution. Furthermore, the function of each entry in a confusion matrix with regard to training set sizes is formulated with an exponential form. Thus, the learning curves of a classification metric can be naturally obtained by transforming the functions of a confusion matrix in terms of the definition of the metric. Moreover, reasonable confidence bands of several popular metrics, including test error, P, R, and F$_1$ score, are derived in this study based on the assumption of the multi-nomial distribution of a confusion matrix. Extensive experiments are conducted on several synthetic and real-world data sets coupled with multiple typical non-neural and neural classification algorithms. Experimental results illustrate the improvements of the proposed learning curves of test error, P, R, and F$_1$ score and the superiority of the confidence bands.

Cite this Paper


BibTeX
@InProceedings{pmlr-v304-xue25a, title = {Learning Curves of Classification Metrics based on Confusion Matrices}, author = {Xue, Yan and Wang, Ruibo and Cao, Xuefei and Yang, Jing and Li, Jihong}, booktitle = {Proceedings of the 17th Asian Conference on Machine Learning}, pages = {1246--1261}, year = {2025}, editor = {Lee, Hung-yi and Liu, Tongliang}, volume = {304}, series = {Proceedings of Machine Learning Research}, month = {09--12 Dec}, publisher = {PMLR}, pdf = {https://raw.githubusercontent.com/mlresearch/v304/main/assets/xue25a/xue25a.pdf}, url = {https://proceedings.mlr.press/v304/xue25a.html}, abstract = {Learning curves of classification metrics, including test error, precision (P), recall (R), F$_1$ score, with regard to training set sizes are a recent hot topic in developing an advanced methodology of model selection and hyperparameter optimization. The existing studies concentrated on formulating the functional shapes of the well-behaved learning curves of test error by using a normality assumption. However, the normality assumption is unreasonable for learning curves of classification metrics because the distributions of most classification metrics, such as P, R, and F$_1$ score, are skewed, and interval estimations of the metrics based on the normality assumption may exceed [0,1]. In this study, considering most classification metrics are obtained from confusion matrices, we develop a novel method to formulate the learning curves of classification metrics by considering that the four entries in a confusion matrix jointly follow a multi-nomial distribution rather than a normality distribution. Furthermore, the function of each entry in a confusion matrix with regard to training set sizes is formulated with an exponential form. Thus, the learning curves of a classification metric can be naturally obtained by transforming the functions of a confusion matrix in terms of the definition of the metric. Moreover, reasonable confidence bands of several popular metrics, including test error, P, R, and F$_1$ score, are derived in this study based on the assumption of the multi-nomial distribution of a confusion matrix. Extensive experiments are conducted on several synthetic and real-world data sets coupled with multiple typical non-neural and neural classification algorithms. Experimental results illustrate the improvements of the proposed learning curves of test error, P, R, and F$_1$ score and the superiority of the confidence bands.} }
Endnote
%0 Conference Paper %T Learning Curves of Classification Metrics based on Confusion Matrices %A Yan Xue %A Ruibo Wang %A Xuefei Cao %A Jing Yang %A Jihong Li %B Proceedings of the 17th Asian Conference on Machine Learning %C Proceedings of Machine Learning Research %D 2025 %E Hung-yi Lee %E Tongliang Liu %F pmlr-v304-xue25a %I PMLR %P 1246--1261 %U https://proceedings.mlr.press/v304/xue25a.html %V 304 %X Learning curves of classification metrics, including test error, precision (P), recall (R), F$_1$ score, with regard to training set sizes are a recent hot topic in developing an advanced methodology of model selection and hyperparameter optimization. The existing studies concentrated on formulating the functional shapes of the well-behaved learning curves of test error by using a normality assumption. However, the normality assumption is unreasonable for learning curves of classification metrics because the distributions of most classification metrics, such as P, R, and F$_1$ score, are skewed, and interval estimations of the metrics based on the normality assumption may exceed [0,1]. In this study, considering most classification metrics are obtained from confusion matrices, we develop a novel method to formulate the learning curves of classification metrics by considering that the four entries in a confusion matrix jointly follow a multi-nomial distribution rather than a normality distribution. Furthermore, the function of each entry in a confusion matrix with regard to training set sizes is formulated with an exponential form. Thus, the learning curves of a classification metric can be naturally obtained by transforming the functions of a confusion matrix in terms of the definition of the metric. Moreover, reasonable confidence bands of several popular metrics, including test error, P, R, and F$_1$ score, are derived in this study based on the assumption of the multi-nomial distribution of a confusion matrix. Extensive experiments are conducted on several synthetic and real-world data sets coupled with multiple typical non-neural and neural classification algorithms. Experimental results illustrate the improvements of the proposed learning curves of test error, P, R, and F$_1$ score and the superiority of the confidence bands.
APA
Xue, Y., Wang, R., Cao, X., Yang, J. & Li, J.. (2025). Learning Curves of Classification Metrics based on Confusion Matrices. Proceedings of the 17th Asian Conference on Machine Learning, in Proceedings of Machine Learning Research 304:1246-1261 Available from https://proceedings.mlr.press/v304/xue25a.html.

Related Material