期刊文献+

基于循环神经网络的汉语语言模型建模方法 被引量:5

Recurrent neural network based Chinese language modeling method
下载PDF
导出
摘要 语言模型是语音识别系统的重要组成部分,目前的主流是n-gram模型。然而n-gram模型存在一些不足,对语句中长距信息描述差、数据稀疏是影响模型性能的两个重要因素。针对不足,研究者提出循环神经网络(Recurrent Neural Network,RNN)建模技术,在英语语言模型建模上取得了较好的效果。根据汉语特点将RNN建模方法应用于汉语语言建模,并结合两种模型的优点,提出了模型融合构建方法。实验结果表明:相比传统的n-gram语言模型,采用RNN训练的汉语语言模型困惑度(Per PLexity,PPL)有了下降,在对汉语电话信道的语音识别上,系统错误率也有下降,将两种语言模型融合后,系统识别错误率更低。 Language model is an important part in the speech recognition system, the current mainstream technique is n-gram model. However, n-gram language model still has some shortcomings: the first is poorly to describe the long-distance information of a sentence, and the second is to arise the serious data sparse phenomenon; essentially they are the two important factors influencing the performances of the model. Aiming at these defects of n-gram language model, the researchers put forward a recurrent neural network(RNN) modeling technique, with which, the training for the English language model has achieved good results. According to the characteristics of the Chinese language, the RNN method is used for training the Chinese language model; also a model combination method to combine the advantages of the two models is proposed. The experimental results show that: the perplexity of RNN model has a certain decline, there is also a certain decline on the system recognition error rate,and after model combination, the recognition error rate reduces much more on the Chinese phone speech recognition, compared with the n-gram language model.
出处 《声学技术》 CSCD 北大核心 2015年第5期431-436,共6页 Technical Acoustics
基金 国家自然科学基金(60872113) 安徽省自然科学基金(1208085MF94 1308085QF99)资助项目
关键词 语音识别 循环神经网络 语言模型 模型融合 speech recognition recurrent neural network language model model combination
  • 相关文献

参考文献15

二级参考文献118

  • 1蔡镜,陈晓,包元晖.基于模糊神经网络的单兵装备效能评估专家系统[J].计算机工程与科学,2005,27(6):98-101. 被引量:4
  • 2钱跃良,林守勋,刘群,刘宏.2005年度863计划中文信息处理与智能人机接口技术评测回顾[J].中文信息学报,2006,20(B03):1-6. 被引量:4
  • 3Zhang, B., S. Matsoukas and R. Schwartz. Discrimina tively trained region dependent teature transforms for speech recognition [C]// Proc. ICASSP, Vol. 1-13, 2006: 313-316. 被引量:1
  • 4Beyerlein, P., et al., Large vocabulary continuous speech recognition of Broadcast News - The Philips/ RWTH approach[J]. Speech Communication, 2002, 37(1-2): 109- 131. 被引量:1
  • 5Hain, T., et al., Automatic transcription of conversational telephone speech [C]// IEEE Transactions on Speech and Audio Processing, 2005, 13(6): 1173-1185. 被引量:1
  • 6Zhang, B. and S. Matsoukas, Minimum phoneme error based heteroscedastic linear discriminant analy sis for speech recognition[C]// Proc. ICASSP, Vol. 1-5, 2005: 1925-1928. 被引量:1
  • 7Hirsimaki, T., et al., Unlimited vocabulary speech recognition with morph language models applied to Finnish[J]. Computer Speech and Language, 2006, 20(4) : 515-541. 被引量:1
  • 8Odell, J.J., The Use of Context in Large Vocabulary Speech Recognition[D]. 1995, University of Cambridge :Cambridge 被引量:1
  • 9Young, S.J., J.J. Odell, and P. C. Woodland. Tree-Based State Tying for High Accuracy Modelling [C]// Proceedings ARPA Workshop on Human Language Technology. 1994. 被引量:1
  • 10Xu, B., et al., Integrating tone information in continuous Mandarin recognition[C]// Proc. ISSPIS, 1999. 被引量:1

共引文献81

同被引文献26

引证文献5

二级引证文献8

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部