跳至主導覽 跳至搜尋 跳過主要內容

SRP-DPCRN-IASDNet: A Blind Sound Source Location Method Based on Deep Neural Networks

研究成果: Article同行評審

摘要

Sound source localization in dynamic environments with multiple moving speakers presents significant challenges due to reverberation, noise, and unknown source counts. To address these issues, this paper proposes an integrated deep-learning framework combining spatial spectrum estimation with blind source detection. The method employs a causal convolution–recurrent network (SRP-DPCRN) to extract robust spatial features from multichannel audio signals under adverse acoustic conditions. Subsequently, an iterative attention-based detection network (IASDNet) automatically identifies active sources from the estimated spatial spectrum without requiring prior knowledge of source quantity. Evaluated on both simulated datasets and the real-recorded LOCATA benchmark, the proposed system demonstrates superior performance in multi-source tracking scenarios, achieving an average detection accuracy of 96% with mean angular error below 3.5 degrees. The results confirm that joint optimization of feature learning and source counting provides an effective solution for blind localization in practical applications, significantly outperforming conventional and deep-learning baselines.

原文English
文章編號698
期刊Mathematics
14
發行號4
DOIs
出版狀態Published - 2月 2026

指紋

深入研究「SRP-DPCRN-IASDNet: A Blind Sound Source Location Method Based on Deep Neural Networks」主題。共同形成了獨特的指紋。

引用此