Attribute-Efficient PAC Learning of Low-Degree Polynomial Threshold Functions with Nasty Noise

Shiwei Zeng, Jie Shen
Proceedings of the 40th International Conference on Machine Learning, PMLR 202:40719-40748, 2023.

Abstract

The concept class of low-degree polynomial threshold functions (PTFs) plays a fundamental role in machine learning. In this paper, we study PAC learning of K-sparse degree-d PTFs on Rn, where any such concept depends only on K out of n attributes of the input. Our main contribution is a new algorithm that runs in time (nd/ϵ)O(d) and under the Gaussian marginal distribution, PAC learns the class up to error rate ϵ with O(K4dϵ2dlog5dn) samples even when an ηO(ϵd) fraction of them are corrupted by the nasty noise of Bshouty et al. (2002), possibly the strongest corruption model. Prior to this work, attribute-efficient robust algorithms are established only for the special case of sparse homogeneous halfspaces. Our key ingredients are: 1) a structural result that translates the attribute sparsity to a sparsity pattern of the Chow vector under the basis of Hermite polynomials, and 2) a novel attribute-efficient robust Chow vector estimation algorithm which uses exclusively a restricted Frobenius norm to either certify a good approximation or to validate a sparsity-induced degree-2d polynomial as a filter to detect corrupted samples.

Cite this Paper


BibTeX
@InProceedings{pmlr-v202-zeng23b, title = {Attribute-Efficient {PAC} Learning of Low-Degree Polynomial Threshold Functions with Nasty Noise}, author = {Zeng, Shiwei and Shen, Jie}, booktitle = {Proceedings of the 40th International Conference on Machine Learning}, pages = {40719--40748}, year = {2023}, editor = {Krause, Andreas and Brunskill, Emma and Cho, Kyunghyun and Engelhardt, Barbara and Sabato, Sivan and Scarlett, Jonathan}, volume = {202}, series = {Proceedings of Machine Learning Research}, month = {23--29 Jul}, publisher = {PMLR}, pdf = {https://proceedings.mlr.press/v202/zeng23b/zeng23b.pdf}, url = {https://proceedings.mlr.press/v202/zeng23b.html}, abstract = {The concept class of low-degree polynomial threshold functions (PTFs) plays a fundamental role in machine learning. In this paper, we study PAC learning of $K$-sparse degree-$d$ PTFs on $\mathbb{R}^n$, where any such concept depends only on $K$ out of $n$ attributes of the input. Our main contribution is a new algorithm that runs in time $({nd}/{\epsilon})^{O(d)}$ and under the Gaussian marginal distribution, PAC learns the class up to error rate $\epsilon$ with $O(\frac{K^{4d}}{\epsilon^{2d}} \cdot \log^{5d} n)$ samples even when an $\eta \leq O(\epsilon^d)$ fraction of them are corrupted by the nasty noise of Bshouty et al. (2002), possibly the strongest corruption model. Prior to this work, attribute-efficient robust algorithms are established only for the special case of sparse homogeneous halfspaces. Our key ingredients are: 1) a structural result that translates the attribute sparsity to a sparsity pattern of the Chow vector under the basis of Hermite polynomials, and 2) a novel attribute-efficient robust Chow vector estimation algorithm which uses exclusively a restricted Frobenius norm to either certify a good approximation or to validate a sparsity-induced degree-$2d$ polynomial as a filter to detect corrupted samples.} }
Endnote
%0 Conference Paper %T Attribute-Efficient PAC Learning of Low-Degree Polynomial Threshold Functions with Nasty Noise %A Shiwei Zeng %A Jie Shen %B Proceedings of the 40th International Conference on Machine Learning %C Proceedings of Machine Learning Research %D 2023 %E Andreas Krause %E Emma Brunskill %E Kyunghyun Cho %E Barbara Engelhardt %E Sivan Sabato %E Jonathan Scarlett %F pmlr-v202-zeng23b %I PMLR %P 40719--40748 %U https://proceedings.mlr.press/v202/zeng23b.html %V 202 %X The concept class of low-degree polynomial threshold functions (PTFs) plays a fundamental role in machine learning. In this paper, we study PAC learning of $K$-sparse degree-$d$ PTFs on $\mathbb{R}^n$, where any such concept depends only on $K$ out of $n$ attributes of the input. Our main contribution is a new algorithm that runs in time $({nd}/{\epsilon})^{O(d)}$ and under the Gaussian marginal distribution, PAC learns the class up to error rate $\epsilon$ with $O(\frac{K^{4d}}{\epsilon^{2d}} \cdot \log^{5d} n)$ samples even when an $\eta \leq O(\epsilon^d)$ fraction of them are corrupted by the nasty noise of Bshouty et al. (2002), possibly the strongest corruption model. Prior to this work, attribute-efficient robust algorithms are established only for the special case of sparse homogeneous halfspaces. Our key ingredients are: 1) a structural result that translates the attribute sparsity to a sparsity pattern of the Chow vector under the basis of Hermite polynomials, and 2) a novel attribute-efficient robust Chow vector estimation algorithm which uses exclusively a restricted Frobenius norm to either certify a good approximation or to validate a sparsity-induced degree-$2d$ polynomial as a filter to detect corrupted samples.
APA
Zeng, S. & Shen, J.. (2023). Attribute-Efficient PAC Learning of Low-Degree Polynomial Threshold Functions with Nasty Noise. Proceedings of the 40th International Conference on Machine Learning, in Proceedings of Machine Learning Research 202:40719-40748 Available from https://proceedings.mlr.press/v202/zeng23b.html.

Related Material