Journal
NEUROCOMPUTING
Volume 431, Issue -, Pages 163-168Publisher
ELSEVIER
DOI: 10.1016/j.neucom.2020.12.048
Keywords
Hypercomplex-valued neural networks; Hyperbolic-valued Hopfield neural networks; Boltzmann machine; Information geometry
Categories
Ask authors/readers for more resources
Information geometry is introduced to analyze hyperbolic-valued neural networks, proving that they form an exponential family and providing natural and mixture parameters, determining the Fisher metric, proving the existence of mixed parameters for all distributions, which are useful for learning algorithms.
Information geometry is a useful tool for analysis of neural networks. In this work, information geometry is introduced to analysis of hyperbolic-valued neural networks. First, hyperbolic-valued Boltzmann machines (HBMs) are organized. Next, the HBMs are analyzed as a neuro manifold from standpoint of information geometry. We prove that the HBMs form an exponential family and provide the natural and mixture parameters. Moreover, the Fisher metric is determined. In addition, we prove the existence of mixed parameters for all the distributions, which are useful for learning algorithms. (C) 2020 Elsevier B.V. All rights reserved.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available