[edit]
Online Self-Concordant and Relatively Smooth Minimization, With Applications to Online Portfolio Selection and Learning Quantum States
Proceedings of The 34th International Conference on Algorithmic Learning Theory, PMLR 201:1481-1483, 2023.
Abstract
Consider an online convex optimization problem where the loss functions are self-concordant barriers, smooth relative to a convex function h, and possibly non-Lipschitz. We analyze the regret of online mirror descent with h. Then, based on the result, we prove the following in a unified manner. Denote by T the time horizon and d the parameter dimension. - For online portfolio selection, the regret of ~EG, a variant of exponentiated gradient due to Helmbold et al., is ˜O(T2/3d1/3) when T>4d/logd. This improves on the original ˜O(T3/4d1/2) regret bound for ~EG. - For online portfolio selection, the regret of online mirror descent with the logarithmic barrier is ˜O(√Td). The regret bound is the same as that of Soft-Bayes due to Orseau et al. up to logarithmic terms. - For online learning quantum states with the logarithmic loss, the regret of online mirror descent with the log-determinant function is also ˜O(√Td). Its per-iteration time is shorter than all existing algorithms we know.