4.5 Article

Deep parameterizations of pairwise and triplet Markov models for unsupervised classification of sequential data

Journal

COMPUTATIONAL STATISTICS & DATA ANALYSIS
Volume 180, Issue -, Pages -

Publisher

ELSEVIER
DOI: 10.1016/j.csda.2022.107663

Keywords

Pairwise Markov chains; Triplet Markov chains; Deep neural networks; Variational expectation -maximization; Image segmentation

Ask authors/readers for more resources

Hidden Markov models are probabilistic graphical models used for classification tasks in time series applications. This paper explores the extensions of these models, such as pairwise and triplet Markov models, which relax the assumptions and introduce new challenges. The paper proposes solutions to address these challenges, including the use of deep neural networks and continuous latent processes.
Hidden Markov models are probabilistic graphical models based on hidden and observed random variables. They are popular to address classification tasks for time series applications such as part-of-speech tagging, image segmentation, genetic sequence analysis. Direct extensions of these models, the pairwise and triplet Markov models, are considered. These models aim at relaxing the assumptions underlying the hidden Markov chain by extending the direct dependencies of the involved random variables or by considering the addition of a third latent process. While these extensions define interesting modeling capabilities that have been little explored so far, they also raise new problems such as defining the nature of their core probability distributions and their parameterization. Once the model is fixed, the unsupervised classification task (i.e. the estimation of the parameters and next of the hidden random variables) is a critical problem. These challenges are addressed, first it is shown that it is possible to embed recent deep neural networks in these models in order to exploit their full modeling power. Second, a continuous latent process in triplet Markov chains is considered. The latter aims at estimating the nature of the joint distributions of the hidden and observed random variables, in addition to their parameters. The introduction of such a continuous auxiliary latent process also offers a new way to model continuous non-stationarity in hidden Markov models. For each model that is introduced, an original unsupervised Bayesian estimation method is proposed. In particular, it takes into account the interpretability of the hidden random variables in terms of signal processing classification. Through unsupervised classification problems on synthetic and real data, it is shown that the new models outperform hidden Markov chains and their classical extensions.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available