International Journal on Advanced Science, Engineering and Information Technology | 2021

Visual Commands for Control of Food Assistance Robot

 
 

Abstract


Assistance robots improve people s quality of life in residential and office tasks, especially for people with physical limitations. In the case of the elderly or people with upper limb motor disabilities, an assistance robot for food support is necessary.\xa0 This development is based on a mixed environment, a real and virtual environment working interactively.\xa0 A camera located in front of the user is used, at a distance of 60 cm, so that it has an excellent visual range to capture the user s hand gestures for the commands. Pattern recognition based on a deep learning algorithm is made with convolutional neural networks to identify the user s hand gestures. This work exposes the network s training and the results of the robot command s execution. A virtual environment is presented in which a robotic arm with a spoon-like effector is used in a machine vision system that allows eight different types of commands to be recognized for the robot by training a faster R-CNN network for which a database of 640 images is used, achieving a degree of system performance of over 95%. The average time in the execution of a cycle from detecting and identify the command gesture to move the robot towards the food and return in front of the user is\xa0 21 seconds, making the development useful for real-time applications.

Volume None
Pages None
DOI 10.18517/ijaseit.11.4.12694
Language English
Journal International Journal on Advanced Science, Engineering and Information Technology

Full Text