摘要
Sound source localization in dynamic environments with multiple moving speakers presents significant challenges due to reverberation, noise, and unknown source counts. To address these issues, this paper proposes an integrated deep-learning framework combining spatial spectrum estimation with blind source detection. The method employs a causal convolution–recurrent network (SRP-DPCRN) to extract robust spatial features from multichannel audio signals under adverse acoustic conditions. Subsequently, an iterative attention-based detection network (IASDNet) automatically identifies active sources from the estimated spatial spectrum without requiring prior knowledge of source quantity. Evaluated on both simulated datasets and the real-recorded LOCATA benchmark, the proposed system demonstrates superior performance in multi-source tracking scenarios, achieving an average detection accuracy of 96% with mean angular error below 3.5 degrees. The results confirm that joint optimization of feature learning and source counting provides an effective solution for blind localization in practical applications, significantly outperforming conventional and deep-learning baselines.
| 原文 | English |
|---|---|
| 文章編號 | 698 |
| 期刊 | Mathematics |
| 卷 | 14 |
| 發行號 | 4 |
| DOIs | |
| 出版狀態 | Published - 2月 2026 |
指紋
深入研究「SRP-DPCRN-IASDNet: A Blind Sound Source Location Method Based on Deep Neural Networks」主題。共同形成了獨特的指紋。引用此
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver