[edit]
Hardness of Agnostically Learning Halfspaces from Worst-Case Lattice Problems
Proceedings of Thirty Sixth Conference on Learning Theory, PMLR 195:3029-3064, 2023.
Abstract
We show hardness of improperly learning halfspaces in the agnostic model, both in the distribution-independent as well as the distribution-specific setting, based on the assumption that worst-case lattice problems, e.g., approximating shortest vectors within polynomial factors, are hard.In particular, we show that under this assumption there is no efficient algorithm that outputs any binary hypothesis, not necessarily a halfspace, achieving misclassfication error better than 12−γ even if the optimal misclassification error is as small is as small as δ.Here, γ can be smaller than the inverse of any polynomial in the dimension and δ as small as exp(−Ω(log1−c(d))), where 0<c<1 is an arbitrary constant and d is the dimension.For the distribution-specific setting, we show that if the marginal distribution is standard Gaussian, for any β>0 learning halfspaces up to error \OPT\LTF+ε takes time at least d˜Ω(1/ε2−β) under the same hardness assumptions.Similarly, we show that learning degree-ℓ polynomial threshold functions up to error \OPT\PTFℓ+\e takes time at least d˜Ω(ℓ2−β/ε4−2β).\OPT\LTF and \OPT\PTFℓ denote the best error achievable by any halfspace or polynomial threshold function, respectively.Our lower bounds qualitively match algorithmic guarantees and (nearly) recover known lower bounds based on non-worst-case assumptions.Previously, such hardness results [D16, DKPZ21] were based on average-case complexity assumptions, specifically, variants of Feige’s random 3SAT hypothesis, or restricted to the statistical query model.Our work gives the first hardness results basing these fundamental learning problems on well-understood worst-case complexity assumption.It is inspired by a sequence of recent works showing hardness of learning well-separated Gaussian mixtures based on worst-case lattice problems.