Hybrid Generative-Discriminative Training of Gaussian Mixture Models
- Master/Diploma Thesis
- Seminar Type
- - None -
- Announcement date
- 11 Oct 2013
- Wolfgang Roth
- Research Areas
Gaussian Mixture Models (GMMs) are a popular choice for modeling probability density functions, with a vast number of applications in speech and audio technologies and machine learning in general. For classification, one can use class-dependent GMMs and easily build a probabilistic classifier using Bayes rule. The problem here is, that the classical way to train GMMs, i.e. using the maximum likelihood principle and applying the expectation-maximization algorithm, is not aware of classification. This generative training typically yields sub-optimal classifiers. On the other hand, when the classifier is trained in a purely discriminative way, the model actually looses most of its probabilistic semantics.
In this master/diploma thesis, the task is to apply a recently proposed hybrid generative-discriminative learning criterion for GMM training. The goal of this approach is to
- yield good classifiers,
- but still allowing generative treatment of the model: marginalization of missing features and semi-supervised learning
You should have
- Good programming skills (Matlab is fine, but C++, Java, Python can also be used)
- Strong interest in Machine Learning (at least absolved CI/EW, or some similar lectures)
- Basic knowledge about probability calculus, ideally having experience with probabilistic graphical models
 C. M. Bishop, Pattern Recognition and Machine Learning, Springer, 2006.
 R. Peharz, S. Tschiatschek, F. Pernkopf, “The Most Generative Maximum Margin Bayesian Networks”, ICML, 2013.
 F. Pernkopf and M. Wohlmayr, “Large Margin Learning of Bayesian Classifiers based on Gaussian Mixture Models”, ECML, 2010