Skip to main content
Article
Comparative Study of Boosting and Non-Boosting Training for Constructing Ensembles of Acoustic Models
Computer Science Department
  • Rong Zhang, Carnegie Mellon University
  • Alexander I Rudnicky, Carnegie Mellon University
Date of Original Version
1-1-2003
Type
Conference Proceeding
Abstract or Description

This paper compares the performance of Boosting and nonBoosting training algorithms in large vocabulary continuous speech recognition (LVCSR) using ensembles of acoustic models. Both algorithms demonstrated significant word error rate reduction on the CMU Communicator corpus. However, both algorithms produced comparable improvements, even though one would expect that the Boosting algorithm, which has a solid theoretic foundation, should work much better than the non-Boosting algorithm. Several voting schemes for hypothesis combining were evaluated, including weighted voting, un-weighted voting and ROVER.

Citation Information
Rong Zhang and Alexander I Rudnicky. "Comparative Study of Boosting and Non-Boosting Training for Constructing Ensembles of Acoustic Models" (2003)
Available at: http://works.bepress.com/alexander_rudnicky/15/