[edit]
Learning Theory for Conditional Risk Minimization
Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, PMLR 54:213-222, 2017.
Abstract
In this work we study the learnability of stochastic processes with respect to the conditional risk, i.e. the existence of a learning algorithm that improves its next-step performance with the amount of observed data. We introduce a notion of pairwise discrepancy between conditional distributions at different times steps and show how certain properties of these discrepancies can be used to construct a successful learning algorithm. Our main results are two theorems that establish criteria for learnability for many classes of stochastic processes, including all special cases studied previously in the literature.