TY - JOUR
T1 - Domain adaptation based on mixture of latent words language models for automatic speech recognition
AU - Masumura, Ryo
AU - Asami, Taichi
AU - Oba, Takanobu
AU - Masataki, Hirokazu
AU - Sakauchi, Sumitaka
AU - Ito, Akinori
N1 - Publisher Copyright:
Copyright © 2018 The Institute of Electronics, Information and Communication Engineers.
PY - 2018/6
Y1 - 2018/6
N2 - This paper proposes a novel domain adaptation method that can utilize out-of-domain text resources and partially domain matched text resources in language modeling. A major problem in domain adaptation is that it is hard to obtain adequate adaptation effects from out-of-domain text resources. To tackle the problem, our idea is to carry out model merger in a latent variable space created from latent words language models (LWLMs). The latent variables in the LWLMs are represented as specific words selected from the observed word space, so LWLMs can share a common latent variable space. It enables us to perform flexible mixture modeling with consideration of the latent variable space. This paper presents two types of mixture modeling, i.e., LWLM mixture models and LWLM cross-mixture models. The LWLM mixture models can perform a latent word space mixture modeling to mitigate domain mismatch problem. Furthermore, in the LWLM cross-mixture models, LMs which individually constructed from partially matched text resources are split into two element models, each of which can be subjected to mixture modeling. For the approaches, this paper also describes methods to optimize mixture weights using a validation data set. Experiments show that the mixture in latent word space can achieve performance improvements for both target domain and out-of-domain compared with that in observed word space.
AB - This paper proposes a novel domain adaptation method that can utilize out-of-domain text resources and partially domain matched text resources in language modeling. A major problem in domain adaptation is that it is hard to obtain adequate adaptation effects from out-of-domain text resources. To tackle the problem, our idea is to carry out model merger in a latent variable space created from latent words language models (LWLMs). The latent variables in the LWLMs are represented as specific words selected from the observed word space, so LWLMs can share a common latent variable space. It enables us to perform flexible mixture modeling with consideration of the latent variable space. This paper presents two types of mixture modeling, i.e., LWLM mixture models and LWLM cross-mixture models. The LWLM mixture models can perform a latent word space mixture modeling to mitigate domain mismatch problem. Furthermore, in the LWLM cross-mixture models, LMs which individually constructed from partially matched text resources are split into two element models, each of which can be subjected to mixture modeling. For the approaches, this paper also describes methods to optimize mixture weights using a validation data set. Experiments show that the mixture in latent word space can achieve performance improvements for both target domain and out-of-domain compared with that in observed word space.
KW - Automatic speech recognition
KW - Domain adaptation
KW - Latent variable space
KW - Latent words language models
KW - Mixture modeling
UR - http://www.scopus.com/inward/record.url?scp=85047996011&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85047996011&partnerID=8YFLogxK
U2 - 10.1587/transinf.2017EDP7210
DO - 10.1587/transinf.2017EDP7210
M3 - Article
AN - SCOPUS:85047996011
SN - 0916-8532
VL - E101D
SP - 1581
EP - 1590
JO - IEICE Transactions on Information and Systems
JF - IEICE Transactions on Information and Systems
IS - 6
ER -