{"title":"Combining multi-distributed mixture models and bayesian networks for semi-supervised learning","authors":"Manuel Stritt, L. Schmidt-Thieme, G. Poeppel","doi":"10.1109/ICMLA.2007.31","DOIUrl":null,"url":null,"abstract":"In many real world scenarios, mixture models have successfully been used for analyzing features in data ([11, 13, 21]). Usually, multivariate Gaussian distributions for continuous data ([2, 8, 4]) or Bayesian networks for nominal data ([15, 16]) are applied. In this paper, we combine both approaches in a family of Bayesian models for continuous data that are able to handle univariate as well as multivariate nodes, different types of distributions, e.g. Gaussian as well as Poisson distributed nodes, and dependencies between nodes. The models we introduce can be used for unsupervised, semi-supervised as well as for fully supervised learning tasks. We evaluate our models empirically on generated synthetic data and on public datasets thereby showing that they outperform classifiers such as SVMs and logistic regression on mixture data.","PeriodicalId":448863,"journal":{"name":"Sixth International Conference on Machine Learning and Applications (ICMLA 2007)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2007-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Sixth International Conference on Machine Learning and Applications (ICMLA 2007)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICMLA.2007.31","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3
Abstract
In many real world scenarios, mixture models have successfully been used for analyzing features in data ([11, 13, 21]). Usually, multivariate Gaussian distributions for continuous data ([2, 8, 4]) or Bayesian networks for nominal data ([15, 16]) are applied. In this paper, we combine both approaches in a family of Bayesian models for continuous data that are able to handle univariate as well as multivariate nodes, different types of distributions, e.g. Gaussian as well as Poisson distributed nodes, and dependencies between nodes. The models we introduce can be used for unsupervised, semi-supervised as well as for fully supervised learning tasks. We evaluate our models empirically on generated synthetic data and on public datasets thereby showing that they outperform classifiers such as SVMs and logistic regression on mixture data.