Online variance minimization

We consider the following type of online variance minimization problem: In every trial t our algorithms get a covariance matrix Ct and try to select a parameter vector wt−1 such that the total variance over a sequence of trials $\sum_{t=1}^{T} (\boldsymbol {w}^{t-1})^{\top} \boldsymbol {C}^{t}\boldsymbol {w}^{t-1}$ is not much larger than the total variance of the best parameter vector u chosen in hindsight. Two parameter spaces in ℝn are considered—the probability simplex and the unit sphere. The first space is associated with the problem of minimizing risk in stock portfolios and the second space leads to an online calculation of the eigenvector with minimum eigenvalue of the total covariance matrix $\sum_{t=1}^{T} \boldsymbol {C}^{t}$. For the first parameter space we apply the Exponentiated Gradient algorithm which is motivated with a relative entropy regularization. In the second case, the algorithm has to maintain uncertainty information over all unit directions u. For this purpose, directions are represented as dyads uu⊤ and the uncertainty over all directions as a mixture of dyads which is a density matrix. The motivating divergence for density matrices is the quantum version of the relative entropy and the resulting algorithm is a special case of the Matrix Exponentiated Gradient algorithm. In each of the two cases we prove bounds on the additional total variance incurred by the online algorithm over the best offline parameter.

[1]  Charles R. Johnson,et al.  Matrix analysis , 1985, Statistical Inference for Engineers and Data Scientists.

[2]  N. Littlestone Learning Quickly When Irrelevant Attributes Abound: A New Linear-Threshold Algorithm , 1987, 28th Annual Symposium on Foundations of Computer Science (sfcs 1987).

[3]  Vladimir Vovk,et al.  Aggregating strategies , 1990, COLT '90.

[4]  N. Littlestone Mistake bounds and logarithmic linear-threshold learning algorithms , 1990 .

[5]  Walter Willinger,et al.  A Nonstandard Approach to Option Pricing , 1991 .

[6]  Manfred K. Warmuth,et al.  The Weighted Majority Algorithm , 1994, Inf. Comput..

[7]  Manfred K. Warmuth,et al.  A comparison of new and old algorithms for a mixture estimation problem , 1995, COLT '95.

[8]  Manfred K. Warmuth,et al.  Additive versus exponentiated gradient updates for linear prediction , 1995, STOC '95.

[9]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.

[10]  T. Cover Universal Portfolios , 1996 .

[11]  Yoram Singer,et al.  On‐Line Portfolio Selection Using Multiplicative Updates , 1998, ICML.

[12]  Manfred K. Warmuth,et al.  Exponentiated Gradient Versus Gradient Descent for Linear Predictors , 1997, Inf. Comput..

[13]  Manfred K. Warmuth,et al.  Averaging Expert Predictions , 1999, EuroCOLT.

[14]  Manfred K. Warmuth,et al.  Relative loss bounds for single neurons , 1999, IEEE Trans. Neural Networks.

[15]  Manfred K. Warmuth,et al.  Tracking a Small Set of Experts by Mixing Past Posteriors , 2003, J. Mach. Learn. Res..

[16]  Mark Herbster,et al.  Tracking the Best Linear Predictor , 2001, J. Mach. Learn. Res..

[17]  Nello Cristianini,et al.  Spectral Kernel Methods for Clustering , 2001, NIPS.

[18]  Santosh S. Vempala,et al.  Efficient algorithms for online decision problems , 2005, J. Comput. Syst. Sci..

[19]  Martin Zinkevich,et al.  Online Convex Programming and Generalized Infinitesimal Gradient Ascent , 2003, ICML.

[20]  Manfred K. Warmuth,et al.  Relative Loss Bounds for Multidimensional Regression Problems , 1997, Machine Learning.

[21]  Mark Herbster,et al.  Tracking the Best Expert , 1995, Machine Learning.

[22]  Gunnar Rätsch,et al.  Matrix Exponentiated Gradient Updates for On-line Learning and Bregman Projection , 2004, J. Mach. Learn. Res..

[23]  Yoram Singer,et al.  A Comparison of New and Old Algorithms for a Mixture Estimation Problem , 1995, COLT '95.

[24]  Dennis S. Bernstein,et al.  Matrix Mathematics: Theory, Facts, and Formulas with Application to Linear Systems Theory , 2005 .

[25]  Sanjeev Arora,et al.  Fast algorithms for approximate semidefinite programming using the multiplicative weights update method , 2005, 46th Annual IEEE Symposium on Foundations of Computer Science (FOCS'05).

[26]  Manfred K. Warmuth,et al.  Optimum Follow the Leader Algorithm , 2005, COLT.

[27]  Manfred K. Warmuth A Bayes Rule for Density Matrices , 2005, NIPS.

[28]  Manfred K. Warmuth,et al.  Online Variance Minimization , 2006, COLT.

[29]  Yoram Singer,et al.  Convex Repeated Games and Fenchel Duality , 2006, NIPS.

[30]  Manfred K. Warmuth,et al.  A Bayesian Probability Calculus for Density Matrices , 2006, UAI.

[31]  Manfred K. Warmuth,et al.  Randomized PCA Algorithms with Regret Bounds that are Logarithmic in the Dimension , 2006, NIPS.

[32]  Stephen P. Boyd,et al.  Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.

[33]  Geoffrey J. Gordon No-regret Algorithms for Online Convex Programs , 2006, NIPS.

[34]  Babak Hassibi,et al.  The p-norm generalization of the LMS algorithm for adaptive filtering , 2003, IEEE Transactions on Signal Processing.

[35]  Gábor Lugosi,et al.  Prediction, learning, and games , 2006 .

[36]  Yishay Mansour,et al.  Improved second-order bounds for prediction with expert advice , 2006, Machine Learning.

[37]  Robert E. Schapire,et al.  Algorithms for portfolio management based on the Newton method , 2006, ICML.

[38]  Manfred K. Warmuth,et al.  Online kernel PCA with entropic matrix updates , 2007, ICML '07.

[39]  Sanjeev Arora,et al.  A combinatorial, primal-dual approach to semidefinite programs , 2007, STOC '07.

[40]  Thierry Paul,et al.  Quantum computation and quantum information , 2007, Mathematical Structures in Computer Science.

[41]  Manfred K. Warmuth,et al.  Learning Permutations with Exponential Weights , 2007, COLT.

[42]  Manfred K. Warmuth Winnowing subspaces , 2007, ICML '07.

[43]  Elad Hazan,et al.  Logarithmic regret algorithms for online convex optimization , 2006, Machine Learning.

[44]  Manfred K. Warmuth When Is There a Free Matrix Lunch? , 2007, COLT.

[45]  Manfred K. Warmuth,et al.  Randomized Online PCA Algorithms with Regret Bounds that are Logarithmic in the Dimension , 2008 .

[46]  A. Moore,et al.  Forecasting Web Page Views: Methods and Observations , 2008 .

[47]  Manfred K. Warmuth,et al.  When Random Play is Optimal Against an Adversary , 2008, COLT.

[48]  Manfred K. Warmuth,et al.  Bayesian generalized probability calculus for density matrices , 2009, Machine Learning.

[49]  Manfred K. Warmuth,et al.  On-line Variance Minimization in O(n2) per Trial? , 2010, COLT.

[50]  Rahul Jain,et al.  QIP = PSPACE , 2011, JACM.

[51]  Ambuj Tewari,et al.  Regularization Techniques for Learning with Matrices , 2009, J. Mach. Learn. Res..