Soft-Bayes: Prod for Mixtures of Experts with Log-Loss

Laurent Orseau, Tor Lattimore, Shane Legg
Proceedings of the 28th International Conference on Algorithmic Learning Theory, PMLR 76:372-399, 2017.

Abstract

We consider prediction with expert advice under the log-loss with the goal of deriving efficient and robust algorithms. We argue that existing algorithms such as exponentiated gradient, online gradient descent and online Newton step do not adequately satisfy both requirements. Our main contribution is an analysis of the Prod algorithm that is robust to any data sequence and runs in linear time relative to the number of experts in each round. Despite the unbounded nature of the log-loss, we derive a bound that is independent of the largest loss and of the largest gradient, and depends only on the number of experts and the time horizon. Furthermore we give a Bayesian interpretation of Prod and adapt the algorithm to derive a tracking regret.

Cite this Paper


BibTeX
@InProceedings{pmlr-v76-orseau17a, title = {Soft-Bayes: Prod for Mixtures of Experts with Log-Loss}, author = {Orseau, Laurent and Lattimore, Tor and Legg, Shane}, booktitle = {Proceedings of the 28th International Conference on Algorithmic Learning Theory}, pages = {372--399}, year = {2017}, editor = {Hanneke, Steve and Reyzin, Lev}, volume = {76}, series = {Proceedings of Machine Learning Research}, month = {15--17 Oct}, publisher = {PMLR}, pdf = {http://proceedings.mlr.press/v76/orseau17a/orseau17a.pdf}, url = {https://proceedings.mlr.press/v76/orseau17a.html}, abstract = {We consider prediction with expert advice under the log-loss with the goal of deriving efficient and robust algorithms. We argue that existing algorithms such as exponentiated gradient, online gradient descent and online Newton step do not adequately satisfy both requirements. Our main contribution is an analysis of the Prod algorithm that is robust to any data sequence and runs in linear time relative to the number of experts in each round. Despite the unbounded nature of the log-loss, we derive a bound that is independent of the largest loss and of the largest gradient, and depends only on the number of experts and the time horizon. Furthermore we give a Bayesian interpretation of Prod and adapt the algorithm to derive a tracking regret.} }
Endnote
%0 Conference Paper %T Soft-Bayes: Prod for Mixtures of Experts with Log-Loss %A Laurent Orseau %A Tor Lattimore %A Shane Legg %B Proceedings of the 28th International Conference on Algorithmic Learning Theory %C Proceedings of Machine Learning Research %D 2017 %E Steve Hanneke %E Lev Reyzin %F pmlr-v76-orseau17a %I PMLR %P 372--399 %U https://proceedings.mlr.press/v76/orseau17a.html %V 76 %X We consider prediction with expert advice under the log-loss with the goal of deriving efficient and robust algorithms. We argue that existing algorithms such as exponentiated gradient, online gradient descent and online Newton step do not adequately satisfy both requirements. Our main contribution is an analysis of the Prod algorithm that is robust to any data sequence and runs in linear time relative to the number of experts in each round. Despite the unbounded nature of the log-loss, we derive a bound that is independent of the largest loss and of the largest gradient, and depends only on the number of experts and the time horizon. Furthermore we give a Bayesian interpretation of Prod and adapt the algorithm to derive a tracking regret.
APA
Orseau, L., Lattimore, T. & Legg, S.. (2017). Soft-Bayes: Prod for Mixtures of Experts with Log-Loss. Proceedings of the 28th International Conference on Algorithmic Learning Theory, in Proceedings of Machine Learning Research 76:372-399 Available from https://proceedings.mlr.press/v76/orseau17a.html.

Related Material