Hartung-Gorre Verlag

Inh.: Dr. Renate Gorre

D-78465 Konstanz

Fon: +49 (0)7533 97227

Fax: +49 (0)7533 97228

www.hartung-gorre.de

S

ETH Series in Information Theory and its Applications, Vol. 10
edited by Amos Lapidoth

 

 

 

Christoph Pfister

 

On Rényi Information Measures

and Their Applications

 

1st edition 2020. X, 174 pages, € 64,00.

ISBN 978-3-86628-663-4

 

 

 

 

 

 

 

 

 

 

 

 

 

Abstract

 

The solutions to many problems in information theory can be expressed using Shannon’s information measures such as entropy, relative entropy, and mutual information. Other problems—for example source coding with exponential cost functions, guessing, and task encoding—require Rényi’s information measures, which generalize Shannon’s.

The contributions of this thesis are as follows: new problems related to guessing, task encoding, hypothesis testing, and horse betting are solved; and two new Rényi measures of dependence and a new conditional Rényi divergence appearing in these problems are analyzed.

The two closely related families of Rényi measures of dependence are studied in detail, and it is shown that they share many properties with Shannon’s mutual information, but the data-processing inequality is only satisfied by one of them. The dependence measures are based on the Rényi divergence and the relative α-entropy, respectively.

The new conditional Rényi divergence is compared with the conditional Rényi divergences that appear in the definitions of the dependence measures by Csiszár and Sibson, and the properties of all three are studied with emphasis on their behavior under data processing. In the same way that Csiszár’s and Sibson’s conditional divergence lead to the respective dependence measures, so does the new conditional divergence lead to the first of the new Rényi dependence measures. Moreover, the new conditional divergence is also related to Arimoto’s measures.

The first solved problem is about guessing with distributed encoders: Two dependent sequences are described separately and, based on their descriptions, have to be determined by a guesser. The description rates are characterized for which the expected number of guesses until correct (or, more generally, its ρ-th moment for some positive ρ) can be driven to one as the length of the sequences tends to infinity. The characterization involves the Rényi entropy and the ArimotoRényi conditional entropy.

The related second problem is about distributed task encoding: Two dependent sequences are described separately, and a decoder produces a list of all the sequence pairs that share the given descriptions. The description rates are characterized for which the expected size of this list (or, more generally, its ρ-th moment for some positive ρ) can be driven to one as the length of the sequences tends to infinity. The characterization involves the Rényi entropy and the second of the new Rényi dependence measures.

The third problem is about a hypothesis testing setup where the observation consists of independent and identically distributed samples from either a known joint probability distribution or an unknown product distribution. The achievable error-exponent pairs are established and the Fenchel biconjugate of the error-exponent function is related to the first of the new Rényi dependence measures. Moreover, an example is provided where the error-exponent function is not convex and thus not equal to its Fenchel biconjugate.

The fourth problem is about horse betting, where, instead of Kelly’s expected log-wealth criterion, a more general family of power-mean utility functions is considered. The key role in the analysis is played by the Rényi divergence, and the setting where the gambler has access to side information motivates the new conditional Rényi divergence. The proposed family of utility functions in the context of gambling with side information also provides another operational meaning to the first of the new Rényi dependence measures. Finally, a universal strategy for independent and identically distributed races is presented that asymptotically maximizes the gambler’s utility function without knowing the winning probabilities or the parameter of the utility function.

 

 

 

Keywords: guessing, task encoding, hypothesis testing, horse betting, Shannon’s information measures, Rényi’s information measures, new conditional Rényi divergence, dependence measures by Csiszár and Sibson, Rényi entropy, ArimotoRényi conditional entropy.

 

 

Reihe "ETH Series in Information Theory and its Applications" im Hartung-Gorre Verlag

 

Direkt bestellen bei / to order directly from:

Hartung-Gorre Verlag / D-78465 Konstanz / Germany

Telefon: +49 (0) 7533 97227  Telefax: +49 (0) 7533 97228
http://www.hartung-gorre.de   eMail: verlag@hartung-gorre.de