Yanhua YangCheng DengShangqian GaoWei LiuDapeng TaoXinbo Gao
As the prosperity of low-cost and easy-operating depth cameras, skeleton-based human action recognition has been extensively studied recently. However, most of the existing methods partially consider that all 3D joints of a human skeleton are identical. Actually, these 3D joints exhibit diverse responses to different action classes, and some joint configurations are more discriminative to distinguish a certain action. In this paper, we propose a discriminative multi-instance multitask learning (MIMTL) framework to discover the intrinsic relationship between joint configurations and action classes. First, a set of discriminative and informative joint configurations for the corresponding action class is captured in multi-instance learning model by regarding the action and the joint configurations as a bag and its instances, respectively. Then, a multitask learning model with group structure constraints is exploited to further reveal the intrinsic relationship between the joint configurations and different action classes. We conduct extensive evaluations of MIMTL using three benchmark 3D action recognition datasets. Experimental results show that our proposed MIMTL framework performs favorably compared with several state-of-the-art approaches.
Hongyang LiJun ChenZengmin XuHuafeng ChenRuimin Hu
Guang ChenManuel GiulianiDaniel ClarkeAndre GaschlerAlois Knoll
Zhennan YanYiqiang ZhanZhigang PengShu LiaoYoshihisa ShinagawaShaoting ZhangDimitris MetaxasXiang Sean Zhou
Biyun ShengJun LiFu XiaoQun LiWankou YangJunwei Han
Jia WuShirui PanXingquan ZhuChengqi ZhangXindong Wu