跳至主導覽 跳至搜尋 跳過主要內容

Multi-scale feature fusion for cross-modality person re-identification: the MSJLNet approach

  • Zhixin Tie
  • , Haobiao Fan
  • , Lingbing Tao
  • , Yanbing Chen
  • , Hao Sheng
  • , Wei Ke

研究成果: Article同行評審

摘要

Visible-Infrared person re-identification (VI-ReID) faces significant challenges due to discrepancies between visible and infrared images. Traditional two-stream networks often struggle to preserve semantic guidance from data augmentation as network depth increases. To address this, we propose the Multi-Scale Joint Learning Network (MSJLNet), which employs a novel four-stream architecture to segregate data-augmented branches from original branches, focusing on extracting robust and color-agnostic modal features. An Information Purification Module (IPM) with a channel attention mechanism is designed to dynamically filter noise and suppress redundant color information in the augmented branches. Furthermore, a Joint Semantic Learning Module (JSLM) effectively fuses global detail features with color-agnostic features, improving the model’s discriminative ability. Extensive experiments on the SYSU-MM01 and RegDB datasets demonstrate MSJLNet’s superior performance, achieving 79.94% Rank-1 accuracy and 74.96% mAP on SYSU-MM01, and 93.14% Rank-1 accuracy and 87.22% mAP on RegDB. The proposed approach offers new insights for enhancing cross-modality feature learning. Code is available at https://github.com/1849714926/MSJLNet.

原文English
文章編號146
期刊Visual Computer
42
發行號2
DOIs
出版狀態Published - 1月 2026

指紋

深入研究「Multi-scale feature fusion for cross-modality person re-identification: the MSJLNet approach」主題。共同形成了獨特的指紋。

引用此