On the Entropy Region of Gaussian Random Variables

Given n (discrete or continuous) random variables X_i, the (2^n-1)-dimensional vector obtained by evaluating the joint entropy of all non-empty subsets of {X_1,...,X_n} is called an entropic vector. Determining the region of entropic vectors is an important open problem with many applications in information theory. Recently, it has been shown that the entropy regions for discrete and continuous random variables, though different, can be determined from one another. An important class of continuous random variables are those that are vector-valued and jointly Gaussian. In this paper we give a full characterization of the convex cone of the entropy region of three jointly Gaussian vector-valued random variables and prove that it is the same as the convex cone of three scalar-valued Gaussian random variables and further that it yields the entire entropy region of 3 arbitrary random variables. We further determine the actual entropy region of 3 vector-valued jointly Gaussian random variables through a conjecture. For n≥4 number of random variables, we point out a set of 2^n-1-(n(n+1))/2 minimal necessary and sufficient conditions that 2^n-1 numbers must satisfy in order to correspond to the entropy vector of n scalar jointly Gaussian random variables. This improves on a result of Holtz and Sturmfels which gave a nonminimal set of conditions. These constraints are related to Cayley's hyperdeterminant and hence with an eye towards characterizing the entropy region of jointly Gaussian random variables, we also present some new results in this area. We obtain a new (determinant) formula for the 2x2x2 hyperdeterminant and we also give a new (transparent) proof of the fact that the principal minors of an n x n symmetric matrix satisfy the 2 x 2 x ... x 2 (up to n times) hyperdeterminant relations.

[1]  Radim Lněnička On the tightness of the Zhang-Yeung inequality for Gaussian vectors , 2003, Commun. Inf. Syst..

[2]  Thomas Wolf,et al.  Hyperdeterminants as integrable discrete systems , 2009, 0903.3864.

[3]  M. Lunelli,et al.  Representation of matroids , 2002, math/0202294.

[4]  Babak Hassibi,et al.  On a Construction of Entropic Vectors Using Lattice-Generated Distributions , 2007, 2007 IEEE International Symposium on Information Theory.

[5]  Terence Chan A combinatorial approach to information inequalities , 2001, Commun. Inf. Syst..

[6]  Zhen Zhang,et al.  A non-Shannon-type conditional inequality of information quantities , 1997, IEEE Trans. Inf. Theory.

[7]  Raymond W. Yeung,et al.  A framework for linear information inequalities , 1997, IEEE Trans. Inf. Theory.

[8]  Terence Chan,et al.  Group characterizable entropy functions , 2007, 2007 IEEE International Symposium on Information Theory.

[9]  Charles R. Johnson,et al.  Bounded Ratios of Products of Principal Minors of Positive Definite Matrices , 2008, 0806.2645.

[10]  Randall Dougherty,et al.  Six New Non-Shannon Information Inequalities , 2006, 2006 IEEE International Symposium on Information Theory.

[11]  Mikael Passare,et al.  Amoebas, Monge-Ampère measures, and triangulations of the Newton polytope. , 2004 .

[12]  Arthur Cayley,et al.  The Collected Mathematical Papers: On the Theory of Linear Transformations , 2009 .

[13]  T. Cover,et al.  Determinant inequalities via information theory , 1988 .

[14]  T. H. Chan Balanced information inequalities , 2003, IEEE Trans. Inf. Theory.

[15]  Frantisek Matús,et al.  Infinitely Many Information Inequalities , 2007, 2007 IEEE International Symposium on Information Theory.

[16]  Nikolai K. Vereshchagin,et al.  A new class of non-Shannon-type inequalities for entropies , 2002, Commun. Inf. Syst..

[17]  Frantisek Matús,et al.  Conditional Independences among Four Random Variables II , 1995, Combinatorics, Probability and Computing.

[18]  T. Chan,et al.  Capacity regions for linear and abelian network codes , 2007, 2007 Information Theory and Applications Workshop.

[19]  Babak Hassibi,et al.  The entropy region for three Gaussian random variables , 2008, 2008 IEEE International Symposium on Information Theory.

[20]  Babak Hassibi,et al.  Normalized Entropy Vectors, Network Information Theory and Convex Optimization , 2007, 2007 IEEE Information Theory Workshop on Information Theory for Wireless Networks.

[21]  Michael J. Tsatsomeros,et al.  Principal minors, Part I: A method for computing all the principal minors of a matrix , 2006 .

[22]  Raymond W. Yeung,et al.  On a relation between information inequalities and group theory , 2002, IEEE Trans. Inf. Theory.

[23]  B. Hassibi,et al.  Cayley's hyperdeterminant, the principal minors of a symmetric matrix and the entropy region of 4 Gaussian random variables , 2008, 2008 46th Annual Allerton Conference on Communication, Control, and Computing.

[24]  P. Wilson,et al.  DISCRIMINANTS, RESULTANTS AND MULTIDIMENSIONAL DETERMINANTS (Mathematics: Theory and Applications) , 1996 .

[25]  Zhen Zhang,et al.  The Capacity Region for Multi-source Multi-sink Network Coding , 2007, 2007 IEEE International Symposium on Information Theory.

[26]  Michael J. Tsatsomeros,et al.  Principal minors, Part II: The principal minor assignment problem , 2006 .

[27]  Ho-Leung Chan,et al.  A combinatorial approach to information inequalities , 1999, 1999 Information Theory and Networking Workshop (Cat. No.99EX371).

[28]  J. Luque,et al.  Polynomial invariants of four qubits , 2002, quant-ph/0212069.

[29]  Bernd Sturmfels,et al.  The hyperdeterminant and triangulations of the 4-cube , 2006, Math. Comput..

[30]  Zhen Zhang,et al.  On Characterization of Entropy Function via Information Inequalities , 1998, IEEE Trans. Inf. Theory.

[31]  Frantisek Matús,et al.  On Gaussian conditional independence structures , 2007, Kybernetika.

[32]  Milan Studený,et al.  Conditional Independences among Four Random Variables 1 , 1995, Comb. Probab. Comput..

[33]  Bernd Sturmfels,et al.  Hyperdeterminantal relations among symmetric principal minors , 2006, math/0604374.

[34]  F. Matús,et al.  Two Constructions on Limits of Entropy Functions , 2007, IEEE Transactions on Information Theory.

[35]  I. M. Gelʹfand,et al.  Discriminants, Resultants, and Multidimensional Determinants , 1994 .

[36]  Zhen Zhang On a new non-Shannon type information inequality , 2003, Commun. Inf. Syst..

[37]  Zhen Zhang,et al.  On a new non-Shannon-type information inequality , 2002, Proceedings IEEE International Symposium on Information Theory,.

[38]  Satoru Fujishige,et al.  Polymatroidal Dependence Structure of a Set of Random Variables , 1978, Inf. Control..