IEEE Access | 2021

Markov Chain Monte Carlo-Based Bayesian Inference for Learning Finite and Infinite Inverted Beta-Liouville Mixture Models

 
 
 
 
 
 

Abstract


Recently Inverted Beta-Liouville mixture models have emerged as an efficient paradigm for proportional positive vectors modeling and unsupervised learning. However, little attention has been devoted to investigate these generative models within discriminative classifiers. Our aim here is to reveal the structure of non-Gaussian data by generating new probabilistic SVM kernels from inverted-Beta Liouville mixture models. The inverted Beta-Liouville has a more general covariance structure and a smaller number of parameters than the inverted Dirichlet and generalized inverted Dirichlet, respectively, which makes it more practical and useful. A principled Bayesian learning algorithm is developed to accurately estimate the model’s parameters. To cope with the problem of selecting the optimal number of components, we further propose a nonparametric Bayesian learning algorithm based on an extended infinite mixture model which may have better modelling and clustering capabilities than the finite model for some applications. Finally, the resulting generative model is exploited to build several efficient probabilistic SVM kernels in order to enhance the expected clustering and modeling performance. Through a number of experimental evaluations involving visual scenes classification, text categorization and texture images discrimination, we prove the merits of the proposed work.

Volume 9
Pages 71170-71183
DOI 10.1109/ACCESS.2021.3078670
Language English
Journal IEEE Access

Full Text