Abstract
Sound source localization in dynamic environments with multiple moving speakers presents significant challenges due to reverberation, noise, and unknown source counts. To address these issues, this paper proposes an integrated deep-learning framework combining spatial spectrum estimation with blind source detection. The method employs a causal convolution–recurrent network (SRP-DPCRN) to extract robust spatial features from multichannel audio signals under adverse acoustic conditions. Subsequently, an iterative attention-based detection network (IASDNet) automatically identifies active sources from the estimated spatial spectrum without requiring prior knowledge of source quantity. Evaluated on both simulated datasets and the real-recorded LOCATA benchmark, the proposed system demonstrates superior performance in multi-source tracking scenarios, achieving an average detection accuracy of 96% with mean angular error below 3.5 degrees. The results confirm that joint optimization of feature learning and source counting provides an effective solution for blind localization in practical applications, significantly outperforming conventional and deep-learning baselines.
| Original language | English |
|---|---|
| Article number | 698 |
| Journal | Mathematics |
| Volume | 14 |
| Issue number | 4 |
| DOIs | |
| Publication status | Published - Feb 2026 |
Keywords
- blind source localization
- casual convolution neural network
- dual path recurrent neural network
- iterative sound source detection
- spatial spectrum
- trajectory tracking
Fingerprint
Dive into the research topics of 'SRP-DPCRN-IASDNet: A Blind Sound Source Location Method Based on Deep Neural Networks'. Together they form a unique fingerprint.Press/Media
-
New Mathematics Study Results from Faculty of Applied Sciences Described (SRP-DPCRN-IASDNet: A Blind Sound Source Location Method Based on Deep Neural Networks)
12/03/26
1 item of Media coverage
Press/Media
Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver