期刊文献+

基于DNN的子空间语音增强算法 被引量:2

A Speech Enhancement Method Based on Deep Neural Network and Subspace Algorithm
在线阅读 下载PDF
导出
摘要 针对噪声的随机性和突变性,使得传统算法抑制非平稳噪声比抑制平稳噪声难度增大的问题,提出了一种基于深度神经网络的子空间语音增强算法。该算法利用带噪的语音信号数据训练一组深度神经网络语音生成型模型(DNN训练模型);在测试增强阶段根据噪声估计和DNN模型去除非平稳噪声;最后,通过信号子空间在抑制噪声和减少信号失真上做出较为折中的选择重构语音信号。实验结果表明,基于深度神经网络的子空间语音增强算法对非平稳噪声有非常强的抑制能力,通过STOI和PESQ值反映了在低信噪比下,该算法可以提高增强语音的可懂度。 The random and abrupt nature of noise makes the traditional algorithm more difficult to suppress non-stationary noise than stationary noise.To solve above problems,a speech enhancement method based on Deep Neural Network and subspace algorithm was proposed.First,the algorithm uses speech signal with noise to train a set of speech production model by deep neural network(DNN training model).Second,in the test of the enhancement noise estimation and DNN model are used to remove non-stationary noise.Finally,the signal subspace in suppressing noise and reducing the signal distortion makes a more eclectic choice of speech signal reconstruction.The simulation results show that the improved algorithm based on speech enhancement can remove non-stationary noise strongly and improve the intelligibility of enhanced speech by STOI under low SNR.
出处 《太原理工大学学报》 CAS 北大核心 2016年第5期647-650,679,共5页 Journal of Taiyuan University of Technology
基金 国家自然基金项目资助:基于认知机理的情感语音识别基础研究(61370093) 山西省青年科技研究基金资助项目(2013021016-1) 山西省自然科学基金资助项目(2013011016-1) 校基金团队资助项目(2014TD028 2014TD029)
关键词 语音增强 信号子空间 深度神经网络 非平稳噪声 噪声估计 speech enhancement subspace deep neural network non-stationary noise noise estimation
  • 相关文献

参考文献8

二级参考文献40

  • 1郑宏兴 姚纪欢 张成.MATLAB 5.X工具箱使用技巧与实例[M].武昌:华中科技大学出版社,chapter4,2001..
  • 2飞思科技产品研法中心.MATLAB 6.5辅助小波分析与应用[M].北京:电子工业出版社,chapter6,2003..
  • 3杨福生.小波变换的工程分析与应用[M].北京:科学出版社,2000..
  • 4Ebert T, Banfer O, Nelles O. Multilayer perceptron network with modified sigmoid activation functions [G]. LNCS 6319: Artificial Intelligence and Computational Intelligence. Berlin: Springer Berlin Heidelberg, 2010: 414-421.
  • 5Karlik B, Olgae AV. Performance analysis of various activa- tion functions in generalized MLP architectures of neural net- works [J]. International Journal of Artificial Intelligence and Expert Systems, 2010, 1 (4): 111-122.
  • 6Glorot X, Bordes A, Bengio Y. Domain adaptation for large-scalesentiment classification: A deep learning approach [C] //Proceedings of the 28th International Conference on Machine Learning, 2011.
  • 7Glorot X, Bengio Y. Understanding the difficulty of training deep feedforward neural networks [C] //Proceedings of the International Conference on Artificial Intelligence and Statistics, 2010.
  • 8Graves A, Mohamed AR, Hinton G. Speech recognition with deep recurrent neural networks [C] //IEEE International Con- ference on Acoustics, Speech and Signal Processing, 2013.
  • 9Pradhan 13, Lee S. Regional landslide susceptibility analysis using back-propagation neural network model at Cameron High- land, Malaysia [J]. Landslides, 2010, 7 (1): 13-30.
  • 10Vincent P, Larochelle H, Lajoie I, et al. Stacked denoising autoencoders: Learning useful representations in a deep net- work with a local denoising criterion [J]. The Journal of Ma- chine LearningResearch, 2010, 11; 3371-3408.

共引文献8

同被引文献23

引证文献2

二级引证文献1

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部