Learning rotations with little regret

We describe online algorithms for learning a rotation from pairs of unit vectors in $$\mathbb {R}^n$$Rn. We show that the expected regret of our online algorithm compared to the best fixed rotation chosen offline over T iterations is $$\sqrt{nT}$$nT. We also give a lower bound that proves that this expected regret bound is optimal within a constant factor. This resolves an open problem posed in COLT 2008. Our online algorithm for choosing a rotation matrix is essentially an incremental gradient descent algorithm over the set of all matrices, with specially tailored projections. We also show that any deterministic algorithm for learning rotations has $$\varOmega (T)$$Ω(T) regret in the worst case.

[1]  Grace Wahba,et al.  Problem 65-1: A least squares estimate of satellite attitude , 1966 .

[2]  J. Stuelpnagel,et al.  A Least Squares Estimate of Satellite Attitude (Grace Wahba) , 1966 .

[3]  P. Schönemann,et al.  A generalized solution of the orthogonal procrustes problem , 1966 .

[4]  G. Stewart The Efficient Generation of Random Orthogonal Matrices with an Application to Condition Estimators , 1980 .

[5]  Philip M. Long,et al.  WORST-CASE QUADRATIC LOSS BOUNDS FOR ON-LINE PREDICTION OF LINEAR FUNCTIONS BY GRADIENT DESCENT , 1993 .

[6]  D. Hestenes,et al.  Lie-groups as Spin groups. , 1993 .

[7]  Manfred K. Warmuth,et al.  Additive versus exponentiated gradient updates for linear prediction , 1995, STOC '95.

[8]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.

[9]  Philip M. Long,et al.  Worst-case quadratic loss bounds for prediction using linear functions and gradient descent , 1996, IEEE Trans. Neural Networks.

[10]  Manfred K. Warmuth,et al.  How to use expert advice , 1997, JACM.

[11]  Manfred K. Warmuth,et al.  Exponentiated Gradient Versus Gradient Descent for Linear Predictors , 1997, Inf. Comput..

[12]  Mark Herbster,et al.  Tracking the Best Linear Predictor , 2001, J. Mach. Learn. Res..

[13]  Martin Zinkevich,et al.  Online Convex Programming and Generalized Infinitesimal Gradient Ascent , 2003, ICML.

[14]  Joan Lasenby,et al.  Applications of Conformal Geometric Algebra in Computer Vision and Graphics , 2004, IWMM/GIAE.

[15]  Gunnar Rätsch,et al.  Matrix Exponentiated Gradient Updates for On-line Learning and Bregman Projection , 2004, J. Mach. Learn. Res..

[16]  Santosh S. Vempala,et al.  Efficient algorithms for online decision problems , 2005, Journal of computer and system sciences (Print).

[17]  Gábor Lugosi,et al.  Prediction, learning, and games , 2006 .

[18]  J. Wissel,et al.  On the Best Constants in the Khintchine Inequality , 2007 .

[19]  Manfred K. Warmuth,et al.  Learning Permutations with Exponential Weights , 2007, COLT.

[20]  Visa Koivunen,et al.  Efficient Riemannian algorithms for optimization under unitary matrix constraint , 2008, 2008 IEEE International Conference on Acoustics, Speech and Signal Processing.

[21]  Visa Koivunen,et al.  Steepest Descent Algorithms for Optimization Under Unitary Matrix Constraint , 2008, IEEE Transactions on Signal Processing.

[22]  Raman Arora,et al.  On Learning Rotations , 2009, NIPS.

[23]  D. Bernstein Matrix Mathematics: Theory, Facts, and Formulas , 2009 .

[24]  Manfred K. Warmuth,et al.  Corrigendum to "Learning rotations with little regret" September 7, 2010 , 2010 .

[25]  Manfred K. Warmuth,et al.  On-line Variance Minimization in O(n2) per Trial? , 2010, COLT.

[26]  Manfred K. Warmuth,et al.  Learning Rotations Online , 2010 .

[27]  Wojciech Kotlowski,et al.  Minimax Algorithm for Learning Rotations , 2011, COLT.

[28]  Manfred K. Warmuth,et al.  Online variance minimization , 2011, Machine Learning.

[29]  Wouter M. Koolen,et al.  Combining initial segments of lists , 2011, Theor. Comput. Sci..