Intell. Serv. Robotics | 2021

Combining CNN and LSTM for activity of daily living recognition with a 3D matrix skeleton representation

 
 

Abstract


In socially assistive robotics, human activity recognition plays a central role when the adaptation of the robot behavior to the human one is required. In this paper, we present an activity recognition approach for activities of daily living based on deep learning and skeleton data. In the literature, ad hoc features extraction/selection algorithms with supervised classification methods have been deployed, reaching an excellent classification performance. Here, we propose a deep learning approach, combining CNN and LSTM, that exploits both the learning of spatial dependencies correlating the limbs in a skeleton 3D grid representation and the learning of temporal dependencies from instances with a periodic pattern that works on raw data and so without requiring an explicit feature extraction process. These models are proposed for real-time activity recognition, and they are tested on the CAD-60 dataset. Results show that the proposed model behaves better than an LSTM model thanks to the automatic features extraction of the limbs’ correlation. “New Person” results show that the CNN-LSTM model achieves $$95.4\\%$$\n \n 95.4\n %\n \n of precision and $$94.4\\%$$\n \n 94.4\n %\n \n of recall, while the “Have Seen” results are $$96.1\\%$$\n \n 96.1\n %\n \n of precision and $$94.7\\%$$\n \n 94.7\n %\n \n of recall.

Volume 14
Pages 175-185
DOI 10.1007/S11370-021-00358-7
Language English
Journal Intell. Serv. Robotics

Full Text