Separation of Reverberant Speech Based on Computational Auditory Scene Analysis
- Авторлар: Hongyan L.1, Meng C.1, Yue W.1
-
Мекемелер:
- College of Information Engineering, Taiyuan University of Technology Taiyuan
- Шығарылым: Том 52, № 6 (2018)
- Беттер: 561-571
- Бөлім: Article
- URL: https://journals.rcsi.science/0146-4116/article/view/175580
- DOI: https://doi.org/10.3103/S0146411618060068
- ID: 175580
Дәйексөз келтіру
Аннотация
This paper proposes a computational auditory scene analysis approach to separation of room reverberant speech, which performs multi-pitch tracking and supervised classification. The algorithm trains speech and non-speech model separately, which learns to map from harmonic features to grouping cue encoding the posterior probability of time-frequency unit being dominated by the target and periodic interference. Then, a likelihood ratio test selects the correct model for labeling time-frequency unit. Experimental results show that the proposed approach produces strong pitch tracking results and leads to significant improvements of predicted speech intelligibility and quality. Compared with the classical Jin-Wang algorithm, the average SNR of this algorithm is improved by 1.22 dB.
Авторлар туралы
Li Hongyan
College of Information Engineering, Taiyuan University of Technology Taiyuan
Хат алмасуға жауапты Автор.
Email: tylihy@163.com
ҚХР, Taiyuan, 030024
Cao Meng
College of Information Engineering, Taiyuan University of Technology Taiyuan
Email: tylihy@163.com
ҚХР, Taiyuan, 030024
Wang Yue
College of Information Engineering, Taiyuan University of Technology Taiyuan
Email: tylihy@163.com
ҚХР, Taiyuan, 030024
Қосымша файлдар
