Administrator by China Associction for Science and Technology
Sponsored by China Society of Automotive Engineers
Published by AUTO FAN Magazine Co. Ltd.

Automotive Engineering ›› 2023, Vol. 45 ›› Issue (12): 2280-2290.doi: 10.19562/j.chinasae.qcgc.2023.12.010

Special Issue: 智能网联汽车技术专题-感知&HMI&测评2023年

Previous Articles     Next Articles

Research on Visible Light and Infrared Post-Fusion Detection Based on TC-YOLOv7 Algorithm

Linhui Li1,2,Xinliang Zhang1,Yifan Fu1,Jing Lian1,2(),Jiaxu Ma1   

  1. 1.School of Automotive Engineering,Dalian University of Technology,Dalian  116024
    2.Dalian University of Technology,State Key Laboratory of Structural Analysis for Industrial Equipment,Dalian  116024
  • Received:2023-04-22 Revised:2023-05-25 Online:2023-12-25 Published:2023-12-21
  • Contact: Jing Lian E-mail:lianjingdlut@126.com

Abstract:

For the problem that it is difficult to achieve fast and accurate detection of visual targets in complex scenes of autonomous driving, a TC-YOLOv7 detection algorithm based on attention mechanism is proposed, which is applied to visible light, infrared and post-fusion scenarios. Firstly, the YOLOv7 benchmark detection model is improved based on the CBAM and Transformer attention mechanism modules, and the performance of visible light and infrared detection is verified by multi-scene datasets. Secondly, the detection methods of three different non-maximum suppression post-fusion methods including SS-PostFusion, DS-PostFusion, and DD-PostFusion are constructed, with the performance verified. Finally, the method combining TC-YOLOv7 and DD-PostFusion is compared with the single-sensor detection results. The results show that the TC-YOLOv7 method has more than 3% accuracy improvement compared with the benchmark method YOLOv7 mAP@.5 in daytime, night, haze, rain, snow visible light and infrared scenes. In the comprehensive scene test set, the TC-YOLOv7 post-fusion method improves the detection accuracy by 4.5% compared with visual light detection, by 11.1% compared with infrared detection and by 0.6% compared with the YOLOv7 post-fusion method. Furthermore, the TC-YOLOv7 post-fusion method inference speed is 39 fps, meeting the real-time requirements of autonomous driving scenarios.

Key words: deep learning, sensor fusion, YOLO, attention mechanism, non-maximum suppression