TY - JOUR
T1 - LogicMix
T2 - Sample mixing data augmentation for multi-label image classification with partial labels
AU - Chong, Chak Fong
AU - Guo, Jielong
AU - Yang, Xu
AU - Ke, Wei
AU - Abreu, Pedro Henriques
AU - Wang, Yapeng
AU - Im, Sio Kei
N1 - Publisher Copyright:
© 2025 Elsevier Ltd
PY - 2026/3
Y1 - 2026/3
N2 - Multi-label image classification datasets are often partially labeled where many labels are missing, posing a significant challenge to training accurate deep classifiers. Most existing approaches assume the missing labels as negatives and/or exploit image and category relationships to regularize training. Orthogonally, this paper studies blending samples in such incomplete datasets as new samples, extending the training data magnitude to increase generalization. First, the proposed LogicMix mixes multiple partially labeled samples to produce new samples, where their unknown labels are naturally mixed by OR's logical equivalences, without replacement with constants. Subsequently, a Decouple Partial-Asymmetric Loss is proposed to assign separate label-focusing policies to original and new samples, addressing the learning imbalance from the different positive-negative label imbalances between original and augmented samples. Finally, we propose a complete learning framework called 2WayAug-PL. LogicMix and conventional data augmentation collaborate to extend the diversity of new samples in both the sample-sample relation and human prior knowledge, while pseudo-labeling compensates for the lack of labels to provide more supervision signals. 27 partially labeled dataset scenarios derived from three benchmarking datasets with various learning difficulties are utilized for comprehensive experiments. LogicMix has shown remarkable effectiveness and generality in improving mAP against compared sample-mixing data augmentation methods. In particular, 2WayAug-PL achieves state-of-the-art average mAP of 84.3 %, 50.1 %, and 93.8 % on MS-COCO, VG-200, and Pascal VOC 2007, respectively. It further pushes the previous best performance achieved by different frameworks by 0.6 % (CFT), 0.6 % (CFT), and 0.1 % (SR). Moreover, 2WayAug-PL significantly outperforms all compared frameworks, as shown by statistical tests. Code is available at: https://github.com/maxium0526/logic_mix.
AB - Multi-label image classification datasets are often partially labeled where many labels are missing, posing a significant challenge to training accurate deep classifiers. Most existing approaches assume the missing labels as negatives and/or exploit image and category relationships to regularize training. Orthogonally, this paper studies blending samples in such incomplete datasets as new samples, extending the training data magnitude to increase generalization. First, the proposed LogicMix mixes multiple partially labeled samples to produce new samples, where their unknown labels are naturally mixed by OR's logical equivalences, without replacement with constants. Subsequently, a Decouple Partial-Asymmetric Loss is proposed to assign separate label-focusing policies to original and new samples, addressing the learning imbalance from the different positive-negative label imbalances between original and augmented samples. Finally, we propose a complete learning framework called 2WayAug-PL. LogicMix and conventional data augmentation collaborate to extend the diversity of new samples in both the sample-sample relation and human prior knowledge, while pseudo-labeling compensates for the lack of labels to provide more supervision signals. 27 partially labeled dataset scenarios derived from three benchmarking datasets with various learning difficulties are utilized for comprehensive experiments. LogicMix has shown remarkable effectiveness and generality in improving mAP against compared sample-mixing data augmentation methods. In particular, 2WayAug-PL achieves state-of-the-art average mAP of 84.3 %, 50.1 %, and 93.8 % on MS-COCO, VG-200, and Pascal VOC 2007, respectively. It further pushes the previous best performance achieved by different frameworks by 0.6 % (CFT), 0.6 % (CFT), and 0.1 % (SR). Moreover, 2WayAug-PL significantly outperforms all compared frameworks, as shown by statistical tests. Code is available at: https://github.com/maxium0526/logic_mix.
KW - Data augmentation
KW - Image recognition
KW - Missing data
KW - Weak supervision
UR - https://www.scopus.com/pages/publications/105012102459
U2 - 10.1016/j.patcog.2025.112186
DO - 10.1016/j.patcog.2025.112186
M3 - Article
AN - SCOPUS:105012102459
SN - 0031-3203
VL - 171
JO - Pattern Recognition
JF - Pattern Recognition
M1 - 112186
ER -