LogicMix: Sample mixing data augmentation for multi-label image classification with partial labels

Chak Fong Chong, Jielong Guo, Xu Yang, Wei Ke, Pedro Henriques Abreu, Yapeng Wang, Sio Kei Im

Research output: Contribution to journalArticlepeer-review

Abstract

Multi-label image classification datasets are often partially labeled where many labels are missing, posing a significant challenge to training accurate deep classifiers. Most existing approaches assume the missing labels as negatives and/or exploit image and category relationships to regularize training. Orthogonally, this paper studies blending samples in such incomplete datasets as new samples, extending the training data magnitude to increase generalization. First, the proposed LogicMix mixes multiple partially labeled samples to produce new samples, where their unknown labels are naturally mixed by OR's logical equivalences, without replacement with constants. Subsequently, a Decouple Partial-Asymmetric Loss is proposed to assign separate label-focusing policies to original and new samples, addressing the learning imbalance from the different positive-negative label imbalances between original and augmented samples. Finally, we propose a complete learning framework called 2WayAug-PL. LogicMix and conventional data augmentation collaborate to extend the diversity of new samples in both the sample-sample relation and human prior knowledge, while pseudo-labeling compensates for the lack of labels to provide more supervision signals. 27 partially labeled dataset scenarios derived from three benchmarking datasets with various learning difficulties are utilized for comprehensive experiments. LogicMix has shown remarkable effectiveness and generality in improving mAP against compared sample-mixing data augmentation methods. In particular, 2WayAug-PL achieves state-of-the-art average mAP of 84.3 %, 50.1 %, and 93.8 % on MS-COCO, VG-200, and Pascal VOC 2007, respectively. It further pushes the previous best performance achieved by different frameworks by 0.6 % (CFT), 0.6 % (CFT), and 0.1 % (SR). Moreover, 2WayAug-PL significantly outperforms all compared frameworks, as shown by statistical tests. Code is available at: https://github.com/maxium0526/logic_mix.

Original languageEnglish
Article number112186
JournalPattern Recognition
Volume171
DOIs
Publication statusPublished - Mar 2026

Keywords

  • Data augmentation
  • Image recognition
  • Missing data
  • Weak supervision

Fingerprint

Dive into the research topics of 'LogicMix: Sample mixing data augmentation for multi-label image classification with partial labels'. Together they form a unique fingerprint.

Cite this