Lower bounds on individual sequence regret

In this work we lower bound the individual sequence anytime regret of a large family of online algorithms. This bound depends on the quadratic variation of the sequence, $$Q_T$$QT, and the learning rate. Nevertheless, we show that any learning rate that guarantees a regret upper bound of $$O(\sqrt{Q_T})$$O(QT) necessarily implies an $$\varOmega (\sqrt{Q_T})$$Ω(QT) anytime regret on any sequence with quadratic variation $$Q_T$$QT. The algorithms we consider are online linear optimization forecasters whose weight vector at time $$t+1$$t+1 is the gradient of a concave potential function of cumulative losses at time t. We show that these algorithms include all linear Regularized Follow the Leader algorithms. We prove our result for the case of potentials with negative definite Hessians, and potentials for the best expert setting satisfying some natural regularity conditions. In the best expert setting, we give our result in terms of the translation-invariant relative quadratic variation. We apply our lower bounds to Randomized Weighted Majority and to linear cost Online Gradient Descent. We show that our analysis can be generalized to accommodate diverse measures of variation beside quadratic variation. We apply this generalized analysis to Online Gradient Descent with a regret upper bound that depends on the variance of losses.

[1]  Yishay Mansour,et al.  Regret to the best vs. regret to the average , 2007, Machine Learning.

[2]  Yurii Nesterov,et al.  Introductory Lectures on Convex Optimization - A Basic Course , 2014, Applied Optimization.

[3]  Sanjeev Arora,et al.  Efficient algorithms for online convex optimization and their applications , 2006 .

[4]  Shai Shalev-Shwartz,et al.  Online learning: theory, algorithms and applications (למידה מקוונת.) , 2007 .

[5]  Yishay Mansour,et al.  Machine Learning Algorithms with Applications in Finance , 2014 .

[6]  Yishay Mansour,et al.  Lower Bounds on Individual Sequence Regret , 2012, ALT.

[7]  Stephen J. Wright,et al.  Optimization for Machine Learning , 2013 .

[8]  Martin Zinkevich,et al.  Online Convex Programming and Generalized Infinitesimal Gradient Ascent , 2003, ICML.

[9]  Stephen P. Boyd,et al.  Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.

[10]  Patrick Jaillet,et al.  Online Optimization , 2011 .

[11]  R. Rockafellar Convex Analysis: (pms-28) , 1970 .

[12]  丸山 徹 Convex Analysisの二,三の進展について , 1977 .

[13]  Eyal Gofer Higher-Order Regret Bounds with Switching Costs , 2014, COLT.

[14]  Yishay Mansour,et al.  Improved second-order bounds for prediction with expert advice , 2006, Machine Learning.

[15]  J. Schur Zwei Sätze über algebraische Gleichungen mit lauter reellen Wurzeln. , .

[16]  Yishay Mansour,et al.  Pricing Exotic Derivatives Using Regret Minimization , 2011, SAGT.

[17]  Santosh S. Vempala,et al.  Efficient algorithms for online decision problems , 2005, Journal of computer and system sciences (Print).

[18]  Elad Hazan The convex optimization approach to regret minimization , 2011 .

[19]  Rong Jin,et al.  25th Annual Conference on Learning Theory Online Optimization with Gradual Variations , 2022 .

[20]  Jean-Yves Audibert Optimization for Machine Learning , 1995 .

[21]  Elad Hazan,et al.  Extracting certainty from uncertainty: regret bounded by variation in costs , 2008, Machine Learning.

[22]  Gábor Lugosi,et al.  Prediction, learning, and games , 2006 .

[23]  Yishay Mansour,et al.  Online trading algorithms and robust option pricing , 2006, STOC '06.