Inh.: Dr. Renate Gorre
Fon: +49 (0)7533 97227
Fax: +49 (0)7533 97228
ETH Series in Information Theory and its Applications,
edited by Amos Lapidoth
On Rényi Information Measures
and Their Applications
1st edition 2020. X, 174 pages, € 64,00.
The solutions to many problems in information theory can be expressed using Shannon’s information measures such as entropy, relative entropy, and mutual information. Other problems—for example source coding with exponential cost functions, guessing, and task encoding—require Rényi’s information measures, which generalize Shannon’s.
The contributions of this thesis are as follows: new problems related to guessing, task encoding, hypothesis testing, and horse betting are solved; and two new Rényi measures of dependence and a new conditional Rényi divergence appearing in these problems are analyzed.
The two closely related families of Rényi measures of dependence are studied in detail, and it is shown that they share many properties with Shannon’s mutual information, but the data-processing inequality is only satisﬁed by one of them. The dependence measures are based on the Rényi divergence and the relative α-entropy, respectively.
The new conditional Rényi divergence is compared with the conditional Rényi divergences that appear in the deﬁnitions of the dependence measures by Csiszár and Sibson, and the properties of all three are studied with emphasis on their behavior under data processing. In the same way that Csiszár’s and Sibson’s conditional divergence lead to the respective dependence measures, so does the new conditional divergence lead to the ﬁrst of the new Rényi dependence measures. Moreover, the new conditional divergence is also related to Arimoto’s measures.
The ﬁrst solved problem is about guessing with distributed encoders: Two dependent sequences are described separately and, based on their descriptions, have to be determined by a guesser. The description rates are characterized for which the expected number of guesses until correct (or, more generally, its ρ-th moment for some positive ρ) can be driven to one as the length of the sequences tends to inﬁnity. The characterization involves the Rényi entropy and the Arimoto–Rényi conditional entropy.
The related second problem is about distributed task encoding: Two dependent sequences are described separately, and a decoder produces a list of all the sequence pairs that share the given descriptions. The description rates are characterized for which the expected size of this list (or, more generally, its ρ-th moment for some positive ρ) can be driven to one as the length of the sequences tends to inﬁnity. The characterization involves the Rényi entropy and the second of the new Rényi dependence measures.
The third problem is about a hypothesis testing setup where the observation consists of independent and identically distributed samples from either a known joint probability distribution or an unknown product distribution. The achievable error-exponent pairs are established and the Fenchel biconjugate of the error-exponent function is related to the ﬁrst of the new Rényi dependence measures. Moreover, an example is provided where the error-exponent function is not convex and thus not equal to its Fenchel biconjugate.
The fourth problem is about horse betting, where, instead of Kelly’s expected log-wealth criterion, a more general family of power-mean utility functions is considered. The key role in the analysis is played by the Rényi divergence, and the setting where the gambler has access to side information motivates the new conditional Rényi divergence. The proposed family of utility functions in the context of gambling with side information also provides another operational meaning to the ﬁrst of the new Rényi dependence measures. Finally, a universal strategy for independent and identically distributed races is presented that asymptotically maximizes the gambler’s utility function without knowing the winning probabilities or the parameter of the utility function.
Keywords: guessing, task encoding, hypothesis testing, horse betting, Shannon’s information measures, Rényi’s information measures, new conditional Rényi divergence, dependence measures by Csiszár and Sibson, Rényi entropy, Arimoto–Rényi conditional entropy.
Direkt bestellen bei / to order directly from:
Hartung-Gorre Verlag / D-78465 Konstanz / Germany