Fisher information under restriction of Shannon information in multi-terminal situations

Fisher information generally decreases by summarizing observed data into encoded messages. The present paper studies the amount of Fisher information included in independently summarized messages from correlated information sources; that is, the amount of Fisher information when sequences xN and yN of N independent observations of random variables x and y are encoded (summarized) independently of each other into meassages mX and mY. The problem is to obtain the maximal amount of Fisher information when the size of the summarized data or Shannon message information is limited. The problem is solved in the case of completely compressed symmetric data summarization. An achievable bound is given in the general case. Information geometry, which is a powerful new differential geometrical method applicable to statistics and systems theory, is applied to this problem, proving its usefulness in information theory as well.

[1]  S. Amari Geometrical theory of asymptotic ancillarity and conditional inference , 1982 .

[2]  S. Amari Differential Geometry of Curved Exponential Families-Curvatures and Information Loss , 1982 .

[3]  S. Amari,et al.  Differential geometry of edgeworth expansions in curved exponential family , 1983 .

[4]  S. Amari,et al.  Geometrical theory of higher-order asymptotics of test, interval estimator and conditional inference , 1983, Proceedings of the Royal Society of London. A. Mathematical and Physical Sciences.

[5]  L. L. Campbell,et al.  The relation between information theory and the differential geometry approach to statistics , 1985, Inf. Sci..

[6]  Shun-ichi Amari,et al.  Differential-geometrical methods in statistics , 1985 .

[7]  Rudolf Ahlswede,et al.  Hypothesis testing with communication constraints , 1986, IEEE Trans. Inf. Theory.

[8]  Te Han,et al.  Hypothesis testing with multiterminal data compression , 1987, IEEE Trans. Inf. Theory.

[9]  Shun-ichi Amari,et al.  Differential geometrical theory of statistics , 1987 .

[10]  R. E. Kass,et al.  Chapter 2: Differential Geometrical Theory of Statistics , 1987 .

[11]  S. Amari Data compression and statistical inference , 1987 .

[12]  Toby Berger,et al.  Estimation via compressed information , 1988, IEEE Trans. Inf. Theory.

[13]  Shun-ichi Amari,et al.  Statistical inference under multiterminal rate restrictions: A differential geometric approach , 1989, IEEE Trans. Inf. Theory.

[14]  Rudolf Ahlswede,et al.  On minimax estimation in the presence of side information about remote data , 1990 .