语音增强改进方法总结

1、模型更复杂

  1. Mel frequency power spectrum (MFP) was used for speech enhancement in INTERSPEECH 2013 :https://bio-asplab.citi.sinica.edu.tw/paper/conference/lu2013speech.pdf
  2. Convolutional maxout neural networks for speech separation:https://ieeexplore.ieee.org/document/7394335
  3. Voice conversion using deep Bidirectional Long Short-Term Memory based Recurrent Neural Networks:https://ieeexplore.ieee.org/document/7178896
  4. Convolutional-recurrent neural networks for speech enhancement:https://ieeexplore.ieee.org/document/8462155
  5. K. Tan, D. Wang. A Convolutional Recurrent Neural Network for Real-Time Speech Enhancement[C]//Interspeech. 2018: 3229-3233.:https://www.researchgate.net/publication/325542192_A_Convolutional_Recurrent_Neural_Network_for_Real-Time_Speech_Enhancement

2、训练方法

  1. Y. Xu, J. Du, L. Dai, et al. An experimental study on speech enhancement based on deep neural networks[J]. IEEE Signal processing letters, 2013, 21(1): 65-68.https://ieeexplore.ieee.org/document/6665000
  2. SNR-Aware Convolutional Neural Network Modeling for Speech Enhancement,Interspeech.2016-211:https://www.researchgate.net/publication/307889660_SNR-Aware_Convolutional_Neural_Network_Modeling_for_Speech_Enhancement
  3. Y. Xu, J. Du, Z. Huang, et al. Multi-objective learning and mask-based post-processing for deep neural network based speech enhancement[J]. arXiv preprint arXiv:1703.07172, 2017.https://www.researchgate.net/publication/315489399_Multi-Objective_Learning_and_Mask-Based_Post-Processing_for_Deep_Neural_Network_Based_Speech_Enhancement
  4. Sun, L. , Du, J. , Dai, L. R. , & Lee, C. H. . (2017). Multiple-target deep learning for LSTM-RNN based speech enhancement. Hands-free Speech Communications & Microphone Arrays. IEEE.https://ieeexplore.ieee.org/document/7895577
  5. Z. Wang, D. Wang. Recurrent deep stacking networks for supervised speech separation[C]//2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2017: 71-75.https://ieeexplore.ieee.org/document/7952120/citations

3、结合其他任务

  1. (VAD)[1] Tian, G. , Du, J. , Xu, Y. , Cong, L. , & Lee, C. H. . (2015). Improving Deep Neural Network Based Speech Enhancement in Low SNR Environments. International Conference on Latent Variable Analysis & Signal Separation. Springer, Cham.http://home.ustc.edu.cn/~gtian09/publications/LVA-ICA2015-015-Gao.pdf
  2. (MFCC)C. Liao, Y. Tsao, X. Lu, et al. Incorporating symbolic sequential modeling for speech enhancement[J]. arXiv preprint arXiv:1904.13142, 2019.:https://arxiv.org/pdf/1904.13142.pdf

4、相位

  1. Williamson, D. S. , Wang, Y. , & Wang, D. L. . (2016). Complex ratio masking for joint enhancement of magnitude and phase. IEEE International Conference on Acoustics. IEEE.https://ieeexplore.ieee.org/document/7472673
  2. Fu, S. W. , Yu, T. , Lu, X. , & Kawai, H. . (2018). Raw waveform-based speech enhancement by fully convolutional networks. 2017 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC). IEEE.https://ieeexplore.ieee.org/document/8281993/citations#citations
  3. S. Pascual, A. Bonafonte, J. Serra. SEGAN: Speech enhancement generative adversarial network[J]. arXiv preprint arXiv:1703.09452, 2017.https://www.researchgate.net/publication/315682472_SEGAN_Speech_Enhancement_Generative_Adversarial_Network

5、多尺度特征

  1. Macartney, C. , & Weyde, T. . (2018). Improved speech enhancement with the wave-u-net.https://www.researchgate.net/publication/329266468_Improved_Speech_Enhancement_with_the_Wave-U-Net
  2. Szu-Wei, Fu, Tao-Wei, Wang, Yu, & Tsao, et al. (2018). End-to-end waveform utterance enhancement for direct evaluation metrics optimization by fully convolutional neural networks. IEEE/ACM Transactions on Audio, Speech, and Language Processing.https://ieeexplore.ieee.org/document/8331910
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值