4.6 Article

Game theory, maximum entropy, minimum discrepancy and robust Bayesian decision theory

期刊

ANNALS OF STATISTICS
卷 32, 期 4, 页码 1367-1433

出版社

INST MATHEMATICAL STATISTICS
DOI: 10.1214/009053604000000553

关键词

additive model; Bayes act; Bregman divergence; Brier score; convexity; duality; equalizer rule; exponential family; Gamma-minimax; generalized exponential family; Kullback-Leibler divergence; logarithmic score; maximin; mean-value constraints; minimax; mutual information; Pythagorean property; redundancy-capacity theorem; relative entropy; saddle-point; scoring rule; specific entropy; uncertainty function; zero-one loss

向作者/读者索取更多资源

We describe and develop a close relationship between two problems that have customarily been regarded as distinct: that of maximizing entropy, and that of minimizing worst-case expected loss. Using a formulation grounded in the equilibrium theory of zero-sum games between Decision Maker and Nature, these two problems are shown to be dual to each other, the solution to each providing that to the other. Although Topsoe described this connection for the Shannon entropy over 20 years ago, it does not appear to be widely known even in that important special case. We here generalize this theory to apply to arbitrary decision problems and loss functions. We indicate how an appropriate generalized definition of entropy can be associated with such a problem, and we show that, subject to certain regularity conditions, the above-mentioned duality continues to apply in this extended context. This simultaneously provides a possible rationale for maximizing entropy and a tool for finding robust Bayes acts. We also describe the essential identity between the problem of maximizing entropy and that of minimizing a related discrepancy or divergence between distributions. This leads to an extension, to arbitrary discrepancies, of a well-known minimax theorem for the case of Kullback-Leibler divergence (the redundancy-capacity theorem of information theory). For the important case of families of distributions having certain mean values specified, we develop simple sufficient conditions and methods for identifying the desired solutions. We use this theory to introduce a new concept of generalized exponential family linked to the specific decision problem under consideration, and we demonstrate that this shares many of the properties of standard exponential families. Finally, we show that the existence of an equilibrium in our game can be rephrased in terms of a Pythagorean property of the related divergence, thus generalizing previously announced results for Kullback-Leibler and Bregman divergences.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据