Jian LüTingting HuangBo ZhaoXiaogai ChenJian ZhouKaibing Zhang
The crucial issue in current methods for skeleton-based action recognition: how to comprehensively capture the evolving features of global context information and temporal dynamics, and how to extract discriminative representations from skeleton joints and body parts. To address these issue, this paper proposes a dual-excitation spatial-temporal graph convolution method. The method adopts a pyramid aggregation structure formed through group convolution, resulting in a pyramid channel-split graph convolution module. The objective is to integrate context information of different scales by splitting channels, facilitating the interaction of information with different dimensions between channels, and establishing dependencies between channels. Subsequently, a motion excitation module is introduced, which activates motion-sensitive channels by grouping feature channels and calculating feature differences across the temporal dimension. This approach forces the model to focus on discriminative features with motion changes. Additionally, a dual attention mechanism is proposed to highlight key joints and body parts within the overall skeleton action sequence, leading to a more interpretable representation for diverse action sequences. On the NTU RGB+D 60 dataset, the accuracy of X-Sub and X-View reaches 91.6% and 96.9%, respectively. On the NTU RGB+D 120 dataset, the accuracy for X-Sub and X-Set is 87.5% and 88.5%, respectively, outperforming other methods and highlighting the effectiveness of the proposed approach in this study.
Daoshuai WangDewei LiYaonan GuanGang WangHaibin Shao
Jianyang XieYanda MengYitian ZhaoAnh NguyenXiaoyun YangYalin Zheng
Fanjia LiAichun ZhuYonggang XuRan CuiGang Hua