Administrator by China Associction for Science and Technology
Sponsored by China Society of Automotive Engineers
Published by AUTO FAN Magazine Co. Ltd.

Automotive Engineering ›› 2024, Vol. 46 ›› Issue (1): 1-8.doi: 10.19562/j.chinasae.qcgc.2024.01.001

    Next Articles

Driver Behavior Recognition Based on Multi-scale Skeleton Graph and Local Visual Context Method

Hongyu Hu,Yechen Li,Zhengguang Zhang,You Qu,Lei He(),Zhenhai Gao   

  1. Jilin University,State Key Laboratory of Automotive Simulation and Control,Changchun  130022
  • Received:2023-07-26 Revised:2023-09-09 Online:2024-01-25 Published:2024-01-23
  • Contact: Lei He E-mail:jlu_helei@jlu.edu.cn

Abstract:

Non-driving behavior identification is one of the important ways to improve the safety of driving. The current recognition method based on skeleton sequence and image fusion has the problems of large model calculation and the difficulty of feature fusion. To address the above problems, the skeleton-image based behavior recognition network (SIBBR-Net) is proposed in this paper, which is based on the multi-scale skeleton graph and the local visual context. SIBBR-Net fully extracts motion and appearance features through a graph convolution network based on multi-scale skeleton graphs and a convolutional neural network based on local vision and attention mechanisms, and better balances the relationship between model representation capabilities and model calculation. The feature bidirectional guided learning strategy based on hand motion, an adaptive feature fusion module and an auxiliary loss on the static feature space can guide mutual guidance and updating between motion and appearance features to achieve adaptive fusion. SIBBR-Net is finally tested on the Drive & Act dataset, and the average accuracy is 61.78% for dynamic labels and 80.42% for static labels. The Floating-point Operations per Second (FLOPS) of SIBBR-Net is 25.92G, which is 76.96% lower than that of the optimal method.

Key words: driver behavior recognition, multi-scale skeleton graph, local visual context, multi-model data adaptive fusion