Separation of Reverberant Speech Based on Computational Auditory Scene Analysis


如何引用文章

全文:

开放存取 开放存取
受限制的访问 ##reader.subscriptionAccessGranted##
受限制的访问 订阅存取

详细

This paper proposes a computational auditory scene analysis approach to separation of room reverberant speech, which performs multi-pitch tracking and supervised classification. The algorithm trains speech and non-speech model separately, which learns to map from harmonic features to grouping cue encoding the posterior probability of time-frequency unit being dominated by the target and periodic interference. Then, a likelihood ratio test selects the correct model for labeling time-frequency unit. Experimental results show that the proposed approach produces strong pitch tracking results and leads to significant improvements of predicted speech intelligibility and quality. Compared with the classical Jin-Wang algorithm, the average SNR of this algorithm is improved by 1.22 dB.

作者简介

Li Hongyan

College of Information Engineering, Taiyuan University of Technology Taiyuan

编辑信件的主要联系方式.
Email: tylihy@163.com
中国, Taiyuan, 030024

Cao Meng

College of Information Engineering, Taiyuan University of Technology Taiyuan

Email: tylihy@163.com
中国, Taiyuan, 030024

Wang Yue

College of Information Engineering, Taiyuan University of Technology Taiyuan

Email: tylihy@163.com
中国, Taiyuan, 030024

补充文件

附件文件
动作
1. JATS XML

版权所有 © Allerton Press, Inc., 2018