Find in Library
Search millions of books, articles, and more
Indexed Open Access Databases
Mixture-Based Probabilistic Graphical Models for the Label Ranking Problem
oleh: Enrique G. Rodrigo, Juan C. Alfaro, Juan A. Aledo, José A. Gámez
| Format: | Article |
|---|---|
| Diterbitkan: | MDPI AG 2021-03-01 |
Deskripsi
The goal of the <i>Label Ranking (LR) problem</i> is to learn <i>preference models</i> that predict the preferred ranking of class labels for a given unlabeled instance. Different well-known machine learning algorithms have been adapted to deal with the LR problem. In particular, fine-tuned instance-based algorithms (e.g., k-nearest neighbors) and model-based algorithms (e.g., decision trees) have performed remarkably well in tackling the LR problem. <i>Probabilistic Graphical Models</i> (<i>PGMs</i>, e.g., <i>Bayesian networks</i>) have not been considered to deal with this problem because of the difficulty of modeling permutations in that framework. In this paper, we propose a <i>Hidden Naive Bayes classifier</i> (<i>HNB</i>) to cope with the LR problem. By introducing a hidden variable, we can design a hybrid Bayesian network in which several types of distributions can be combined: multinomial for discrete variables, Gaussian for numerical variables, and <i>Mallows</i> for permutations. We consider two kinds of probabilistic models: one based on a <i>Naive Bayes</i> graphical structure (where only univariate probability distributions are estimated for each state of the hidden variable) and another where we allow interactions among the predictive attributes (using a multivariate Gaussian distribution for the parameter estimation). The experimental evaluation shows that our proposals are competitive with the start-of-the-art algorithms in both accuracy and in CPU time requirements.