Separation of Reverberant Speech Based on Computational Auditory Scene Analysis
- 作者: Hongyan L.1, Meng C.1, Yue W.1
-
隶属关系:
- College of Information Engineering, Taiyuan University of Technology Taiyuan
- 期: 卷 52, 编号 6 (2018)
- 页面: 561-571
- 栏目: Article
- URL: https://journals.rcsi.science/0146-4116/article/view/175580
- DOI: https://doi.org/10.3103/S0146411618060068
- ID: 175580
如何引用文章
详细
This paper proposes a computational auditory scene analysis approach to separation of room reverberant speech, which performs multi-pitch tracking and supervised classification. The algorithm trains speech and non-speech model separately, which learns to map from harmonic features to grouping cue encoding the posterior probability of time-frequency unit being dominated by the target and periodic interference. Then, a likelihood ratio test selects the correct model for labeling time-frequency unit. Experimental results show that the proposed approach produces strong pitch tracking results and leads to significant improvements of predicted speech intelligibility and quality. Compared with the classical Jin-Wang algorithm, the average SNR of this algorithm is improved by 1.22 dB.
作者简介
Li Hongyan
College of Information Engineering, Taiyuan University of Technology Taiyuan
编辑信件的主要联系方式.
Email: tylihy@163.com
中国, Taiyuan, 030024
Cao Meng
College of Information Engineering, Taiyuan University of Technology Taiyuan
Email: tylihy@163.com
中国, Taiyuan, 030024
Wang Yue
College of Information Engineering, Taiyuan University of Technology Taiyuan
Email: tylihy@163.com
中国, Taiyuan, 030024
补充文件
