WiFi-based human gesture recognition has recently enjoyed increasing popularity in the Internet of Things (IoT) scenarios. Simultaneously recognizing user identities and user gestures is of great importance for enhancing the system security and user quality of experience (QoE). State-of-the-art approaches that perform dual tasks suffer from increased latency or degraded accuracy in cross-domain scenarios. In this paper, we present WiDual, a dual-task system that achieves accurate cross-domain gesture recognition and user identification based on WiFi in a real-time manner. The basic idea of WiDual is to use the attention mechanism to adaptively explore cross-domain features worthy of attention for dual tasks. WiDual employs a CSI (Channel Statement Information) visualization method that transfers WiFi signals to images for further feature extraction and model training. In this way, WiDual mitigates the possible loss of useful information and excessive delays caused by extracting handcrafted features directly from the WiFi signal. Furthermore, WiDual utilizes a collaboration module to combine gesture features and user identity features to enhance the performance of dual-task recognition. We implement WiDual and evaluate its performance extensively on a public dataset including 6 gestures and 6 users performed across domains. Results show that WiDual outperforms state-of-the-art approaches, with 26% and 8% improvements on the accuracy of cross-domain user identification and gesture recognition respectively.
Chenning LiManni LiuZhichao Cao
Raghav H. VenkatnarayanGriffin PageMuhammad Shahzad
Raghav H. VenkatnarayanShakir MahmoodMuhammad Shahzad
Yuxi QinStephan SiggSu PanZibo Li
Chenning LiManni LiuZhichao Cao