IEEE Transactions on Instrumentation and Measurement | 2021

Learning Disentangled Representation for Mixed- Reality Human Activity Recognition With a Single IMU Sensor

 
 
 
 
 
 

Abstract


Together with the rapid development of the sensors technology in recent years, sensor-based human activity recognition (HAR) has shown promising performance using well-known supervised deep learning methods. However, it remains challenging in a realistic scenario, i.e., limited number of labeled samples and sensors. This article proposes a novel deep learning method to achieve accurate and robust HAR with only a single inertial measurement unit (IMU) sensor. Our contributions are twofold. First, based on the skinned multiperson linear (SMPL) model, we build a large synthetic HAR dataset containing multimodal measurements: acceleration and angular velocity, which were generated according to the forward kinematics. Second, We propose a multiple-level domain adaptive learning model with information-theoretically stimulated constraints to simultaneously align the distributions of low- and high-level representations of virtual and real HAR data. The proposed mutual information constraints encourage the neural network to learn a disentangled representation for the multimodal sensing data. Comprehensive experimental results on three publicly available datasets demonstrate that the proposed method compares favorably with competing ones and has robust performance with variable labeled samples.

Volume 70
Pages 1-14
DOI 10.1109/TIM.2021.3111996
Language English
Journal IEEE Transactions on Instrumentation and Measurement

Full Text