TY - GEN
T1 - Adaptive Optimization Strategies for Gigapixel Object Detection
AU - Zhang, Runze
AU - Lu, Lu
AU - He, Meng
AU - Fan, Baoyu
AU - Li, Xiaochuan
AU - Guo, Zhenhua
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - GigaPixel-level computer vision tasks recently become new research hotspots, due to the development of photography. Object detection, as a basic, common, but challenging task, undoubtedly received the most attention. However, most of the research focused on the efficiency improvements for the super resolution of the scenarios. They tend to design relevant network modules or inference strategies to help split the whole image into smaller patches for efficient computation. Differently from them, We proposed three optimization strategies that can maintain efficient computation while also ensuring the accuracy of model inference strategies. The strategies are Anchor-Split Sample Strategy, GPU Memory Optimizations and Two-Phase Adaptive Inference Strategy. Anchor-Split Sample Strategy can help train the detectors within 8 hours on the PANDA detection datasets. GPU Memory Optimizations can help train the DETA model with Swin-Large backone on a consumer GPU card like RTX 3080 with only costing 18G memory. Two-Phase Adaptive Inference Strategy, Without the extra training of the additional network modules or complex strategies, can obtain 74% mAP and 82% AR500 performance with only 1.5h cost on the 15W Power Jetson Orin AGX card. Compared with the state-of-the-art methods, our methods can boost the performance by 20% percentage.
AB - GigaPixel-level computer vision tasks recently become new research hotspots, due to the development of photography. Object detection, as a basic, common, but challenging task, undoubtedly received the most attention. However, most of the research focused on the efficiency improvements for the super resolution of the scenarios. They tend to design relevant network modules or inference strategies to help split the whole image into smaller patches for efficient computation. Differently from them, We proposed three optimization strategies that can maintain efficient computation while also ensuring the accuracy of model inference strategies. The strategies are Anchor-Split Sample Strategy, GPU Memory Optimizations and Two-Phase Adaptive Inference Strategy. Anchor-Split Sample Strategy can help train the detectors within 8 hours on the PANDA detection datasets. GPU Memory Optimizations can help train the DETA model with Swin-Large backone on a consumer GPU card like RTX 3080 with only costing 18G memory. Two-Phase Adaptive Inference Strategy, Without the extra training of the additional network modules or complex strategies, can obtain 74% mAP and 82% AR500 performance with only 1.5h cost on the 15W Power Jetson Orin AGX card. Compared with the state-of-the-art methods, our methods can boost the performance by 20% percentage.
KW - GigaPixel
KW - Object detection
KW - Optimization Strategies
UR - http://www.scopus.com/inward/record.url?scp=85184377487&partnerID=8YFLogxK
U2 - 10.1109/PAAP60200.2023.10391523
DO - 10.1109/PAAP60200.2023.10391523
M3 - Conference contribution
AN - SCOPUS:85184377487
T3 - Proceedings - International Symposium on Parallel Architectures, Algorithms and Programming, PAAP
BT - Proceedings - 2023 The 14th International Symposium on Parallel Architectures, Algorithms and Programming, PAAP 2023
PB - IEEE Computer Society
T2 - 14th International Symposium on Parallel Architectures, Algorithms and Programming, PAAP 2023
Y2 - 24 November 2023 through 26 November 2023
ER -