4.5 Article

Information Theory With Kernel Methods

期刊

IEEE TRANSACTIONS ON INFORMATION THEORY
卷 69, 期 2, 页码 752-775

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TIT.2022.3211077

关键词

Kernel; Entropy; Probability distribution; Information theory; Hilbert space; Eigenvalues and eigenfunctions; Tensors; Kernel methods; quantum information theory

向作者/读者索取更多资源

Through the associated covariance operators in reproducing kernel Hilbert spaces, we analyze probability distributions. We find that the von Neumann entropy and relative entropy of these operators are closely related to the usual notions of Shannon entropy and relative entropy, and have many similar properties. They can be used together with efficient estimation algorithms from various oracles on the probability distributions. We also consider product spaces and define notions of mutual information and joint entropies for tensor product kernels, which perfectly characterize independence, but only partially conditional independence. Lastly, we show how these new notions of relative entropy lead to new upper-bounds on log partition functions, which can be utilized in conjunction with convex optimization in variational inference methods, providing a new family of probabilistic inference methods.
We consider the analysis of probability distributions through their associated covariance operators from reproducing kernel Hilbert spaces. We show that the von Neumann entropy and relative entropy of these operators are intimately related to the usual notions of Shannon entropy and relative entropy, and share many of their properties. They come together with efficient estimation algorithms from various oracles on the probability distributions. We also consider product spaces and show that for tensor product kernels, we can define notions of mutual information and joint entropies, which can then characterize independence perfectly, but only partially conditional independence. We finally show how these new notions of relative entropy lead to new upper-bounds on log partition functions, that can be used together with convex optimization within variational inference methods, providing a new family of probabilistic inference methods.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.5
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据