ISCA Archive Interspeech 2013
ISCA Archive Interspeech 2013

Recurrent neural network based language model personalization by social network crowdsourcing

Tsung-Hsien Wen, Aaron Heidel, Hung-yi Lee, Yu Tsao, Lin-shan Lee

Speech recognition has become an important feature in smartphones in recent years. Different from traditional automatic speech recognition, the speech recognition on smartphones can take advantage of personalized language models to model the linguistic patterns and wording habits of a particular smartphone owner better. Owing to the popularity of social networks in recent years, personal texts and messages are no longer inaccessible. However, data sparseness is still an unsolved problem. In this paper, we propose a three-step adaptation approach to personalize recurrent neural network language models (RNNLMs). We believe that its capability to model word histories as distributed representations of arbitrary length can help mitigate the data sparseness problem. Furthermore, we also propose additional user-oriented features to empower the RNNLMs with stronger capabilities for personalization. The experiments on a Facebook dataset showed that the proposed method not only drastically reduced the model perplexity in preliminary experiments, but also moderately reduced the word error rate in n-best rescoring tests.


doi: 10.21437/Interspeech.2013-621

Cite as: Wen, T.-H., Heidel, A., Lee, H.-y., Tsao, Y., Lee, L.-s. (2013) Recurrent neural network based language model personalization by social network crowdsourcing. Proc. Interspeech 2013, 2703-2707, doi: 10.21437/Interspeech.2013-621

@inproceedings{wen13_interspeech,
  author={Tsung-Hsien Wen and Aaron Heidel and Hung-yi Lee and Yu Tsao and Lin-shan Lee},
  title={{Recurrent neural network based language model personalization by social network crowdsourcing}},
  year=2013,
  booktitle={Proc. Interspeech 2013},
  pages={2703--2707},
  doi={10.21437/Interspeech.2013-621}
}