Higher-Order Regret Bounds with Switching Costs
Proceedings of The 27th Conference on Learning Theory, PMLR 35:210-243, 2014.
This work examines online linear optimization with full information and switching costs (SCs) and focuses on regret bounds that depend on properties of the loss sequences. The SCs considered are bounded functions of a pair of decisions, and regret is augmented with the total SC. We show under general conditions that for any normed SC, σ(\mathbfx,\mathbfx’)=\|\mathbfx-\mathbfx’\|, regret \textitcannot be bounded given only a bound Q on the quadratic variation of losses. With an additional bound Λon the total length of losses, we prove O(\sqrtQ+Λ) regret for Regularized Follow the Leader (RFTL). Furthermore, an O(\sqrtQ) bound holds for RFTL given a cost \|\mathbfx-\mathbfx’\|^2. By generalizing the Shrinking Dartboard algorithm, we also show an expected regret bound for the best expert setting with any SC, given bounds on the total loss of the best expert and the quadratic variation of any expert. As SCs vanish, all our bounds depend purely on quadratic variation. We apply our results to pricing options in an arbitrage-free market with proportional transaction costs. In particular, we upper bound the price of “at the money” call options, assuming bounds on the quadratic variation of a stock price and the minimum of summed gains and summed losses.