TY - JOUR
T1 - Round-robin duel discriminative language models
AU - Oba, Takanobul
AU - Hori, Takaaki
AU - Nakamura, Atsushi
AU - Ito, Akinori
N1 - Funding Information:
Manuscript received January 03, 2011; revised May 30, 2011; accepted October 20, 2011. Date of publication October 31, 2011; date of current version February 24, 2012. The research described in this paper was supported in part by the Japan Society for the Promotion of Science under Grant-in-Aid Scientific Research No. 22300064. The associate editor coordinating the review of this manuscript and approving it for publication was Prof. Hui Jiang.
PY - 2012
Y1 - 2012
N2 - Discriminative training has received a lot of attention from both the machine learning and speech recognition communities. The idea behind the discriminative approach is to construct a model that distinguishes correct samples from incorrect samples, while the conventional generative approach estimates the distributions of correct samples. We propose a novel discriminative training method and apply it to a language model for reranking speech recognition hypotheses. Our proposed method has round-robin duel discrimination (R2D2) criteria in which all the pairs of sentence hypotheses including pairs of incorrect sentences are distinguished from each other, taking their error rate into account. Since the objective function is convex, the global optimum can be found through a normal parameter estimation method such as the quasi-Newton method. Furthermore, the proposed method is an expansion of the global conditional log-linear model whose objective function corresponds to the conditional random fields. Our experimental results show that R2D2 outperforms conventional methods in many situations, including different languages, different feature constructions and different difficulties.
AB - Discriminative training has received a lot of attention from both the machine learning and speech recognition communities. The idea behind the discriminative approach is to construct a model that distinguishes correct samples from incorrect samples, while the conventional generative approach estimates the distributions of correct samples. We propose a novel discriminative training method and apply it to a language model for reranking speech recognition hypotheses. Our proposed method has round-robin duel discrimination (R2D2) criteria in which all the pairs of sentence hypotheses including pairs of incorrect sentences are distinguished from each other, taking their error rate into account. Since the objective function is convex, the global optimum can be found through a normal parameter estimation method such as the quasi-Newton method. Furthermore, the proposed method is an expansion of the global conditional log-linear model whose objective function corresponds to the conditional random fields. Our experimental results show that R2D2 outperforms conventional methods in many situations, including different languages, different feature constructions and different difficulties.
KW - Discriminative language model
KW - error correction
KW - round-robin duel discrimination (R2D2)
UR - http://www.scopus.com/inward/record.url?scp=84857478465&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84857478465&partnerID=8YFLogxK
U2 - 10.1109/TASL.2011.2174225
DO - 10.1109/TASL.2011.2174225
M3 - Article
AN - SCOPUS:84857478465
SN - 1558-7916
VL - 20
SP - 1244
EP - 1255
JO - IEEE Transactions on Speech and Audio Processing
JF - IEEE Transactions on Speech and Audio Processing
IS - 4
M1 - 6064876
ER -