期刊文献+

一个语言模型压缩方法的研究与实践 被引量:1

The Research and Practice of a Language Model Compression
下载PDF
导出
摘要 随着手持设备的发展,语言模型压缩在研究中占据着重要位置。使用互信息和熵差相结合的方法对二元模型进行压缩。其基本思想是首先使用互信息对二元信息重要性进行判断,然后使用基于熵差的压缩方法得到最终的语言模型,以困惑度为评价标准将使用该方法压缩后的语言模型与其他方法进行比较。实验结果表明该方法得到的模型性能更好。 With the development of handsets, the research about the language model compression becomes increasingly important. In this paper we compress the bi-gram model in the way of combining the mutual information and the difference of entropy. This method firstly judges the importance of the bi-gram with mutual information. Then it uses the entropy-based pruning to get the final language model. We compare this method with other methods using the perplexity. The results show that the performance of the language model using this method is better.
出处 《苏州大学学报(工科版)》 CAS 2008年第3期16-20,共5页 Journal of Soochow University Engineering Science Edition (Bimonthly)
基金 高等学校博士学科点专项科研基金项目(编号20060285008)
关键词 语言模型压缩 互信息 熵差 困惑度 language model compression mutual information difference of entropy perplexity
  • 相关文献

参考文献6

二级参考文献16

  • 1周强.基于语料库和面向统计学的自然语言处理技术[J].计算机科学,1995,22(4):36-40. 被引量:25
  • 2Chen Stanleyf,博士学位论文,1996年 被引量:1
  • 3Schenand J Goodman. An empirical study of smoothing techniques for language modeling[J]. Proceedings of the 34th meeting of the ACL, 1996:310-318. 被引量:1
  • 4Slava M. Katz. Estimation of probabilities from sparse data for the language model component of a speech recognizer[J]. IEEE transactions on acoustics, speech, and signal processing, March 1987, vol. ASSP-35(3):400-401. 被引量:1
  • 5Fu Zu-yun.Foundations of information theory[M].Beijing:Publishing House of Electronics Industry,1989. 被引量:1
  • 6Gao Jun.The study and application of chinese language models[D].Beijing University of Posts and Telecommunications,1998,5. 被引量:1
  • 7Ronald Rosenfeld.A maximum entropy to adaptive statistical language learning[J].Computer Speech ang Language,1996,10(3):187-228. 被引量:1
  • 8Huang X,Alleva F,Hwang M Y,et al.An overview of the SPHINX-Ⅱ speech recognition system[C].Proc in DARPA Human Language Technology Workshop,Published as Human Language Technology,San Francisco,CA:Morgan Kaufmann.1993,3:81-86. 被引量:1
  • 9Joshua T.Goodman.A bit of progress in language modeling[J].Computer Speech and Language,2001,15(4):403-434. 被引量:1
  • 10Xu Zhi-ming,Wang Xiao-long,Guan Yi.The data smooth technology of N-gram language models[J].Application Research of Computers.1999,16(7):37-39. 被引量:1

共引文献32

同被引文献11

引证文献1

二级引证文献1

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部