Skip to main content
Unpublished Paper
High-Performance Semi-Supervised Learning using Discriminatively Constrained Generative Models
(2010)
  • Gregory Druck
  • Andrew McCallum, University of Massachusetts - Amherst
Abstract
We develop a semi-supervised learning algorithm that encourages generative models to discover latent structure that is relevant to a prediction task. The method constrains the posterior distribution of latent variables under a generative model to satisfy a rich set of feature expectation constraints from labeled data. We focus on the application of this method to sequence labeling and estimate parameters with a modified EM algorithm. The E-step involves estimating the parameters of a log-linear model with an HMM as the base distribution. This HMM-CRF can be used for test time prediction. The approach is related to other semi-supervised methods, but can be written as a single objective function, has convergence guarantees, and affords additional flexibility, for example to use different latent and output spaces. We evaluate the method on three sequence labeling tasks, achieving the best reported results on two of them, and showing promising results on CoNLL03 named-entity recognition.
Disciplines
Publication Date
2010
Comments
This is the pre-published version harvested from CIIR.
Citation Information
Gregory Druck and Andrew McCallum. "High-Performance Semi-Supervised Learning using Discriminatively Constrained Generative Models" (2010)
Available at: http://works.bepress.com/andrew_mccallum/77/