3.8 Proceedings Paper

Easy Variational Inference for Categorical Models via an Independent Binary Approximation

Publisher

JMLR-JOURNAL MACHINE LEARNING RESEARCH

Keywords

-

Funding

  1. U.S. Army DEVCOM Soldier Center [W911QY-19-2-0003]
  2. U.S. National Science Foundation [HDR-1934553]
  3. NSF [IIS-1908617, CCF 1553075, DRL 1931978, EEC 1937057, 1934553, 1935555, 1931978, 1937057]
  4. AFOSR [FA9550-18-1-0465]
  5. Directorate For Engineering
  6. Div Of Engineering Education and Centers [1937057] Funding Source: National Science Foundation
  7. Division of Computing and Communication Foundations
  8. Direct For Computer & Info Scie & Enginr [1934553] Funding Source: National Science Foundation
  9. Division Of Research On Learning
  10. Direct For Education and Human Resources [1931978] Funding Source: National Science Foundation
  11. Div Of Chem, Bioeng, Env, & Transp Sys
  12. Directorate For Engineering [1935555] Funding Source: National Science Foundation

Ask authors/readers for more resources

This paper presents a new class of generalized linear models (GLMs) for categorical data and proposes a posterior approximation method based on binary categorical-from-binary (CB) models, which allows for fast and simple inference. The quality of the approximation can be improved through Bayesian model averaging. Experimental results show that this method outperforms other posterior estimation methods in dealing with a large number of categories.
We pursue tractable Bayesian analysis of generalized linear models (GLMs) for categorical data. GLMs have been difficult to scale to more than a few dozen categories due to non-conjugacy or strong posterior dependencies when using conjugate auxiliary variable methods. We define a new class of GLMs for categorical data called categorical-from-binary (CB) models. Each CB model has a likelihood that is bounded by the product of binary likelihoods, suggesting a natural posterior approximation. This approximation makes inference straight-forward and fast; using well-known auxiliary variables for probit or logistic regression, the product of binary models admits conjugate closedform variational inference that is embarrassingly parallel across categories and invariant to category ordering. Moreover, an independent binary model simultaneously approximates multiple CB models. Bayesian model averaging over these can improve the quality of the approximation for any given dataset. We show that our approach scales to thousands of categories, outperforming posterior estimation competitors like Automatic Differentiation Variational Inference (ADVI) and No U-Turn Sampling (NUTS) in the time required to achieve fixed prediction quality.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

3.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available