Deep Learning Model for Dynamic Hand Gesture Recognition for Natural Human-Machine Interface on End Devices

Deep Learning Model for Dynamic Hand Gesture Recognition for Natural Human-Machine Interface on End Devices

Tsui-Ping Chang, Hung-Ming Chen, Shih-Ying Chen, Wei-Cheng Lin
Copyright: © 2022 |Pages: 23
DOI: 10.4018/IJISMD.306636
Article PDF Download
Open access articles are freely available for download

Abstract

As end devices have become ubiquitous in daily life, the use of natural human-machine interfaces has become an important topic. Many researchers have proposed the frameworks to improve the performance of dynamic hand gesture recognition. Some CNN models are widely used to increase the accuracy of dynamic hand gesture recognition. However, most CNN models are not suitable for end devices. This is because image frames are captured continuously and result in lower hand gesture recognition accuracy. In addition, the trained models need to be efficiently deployed on end devices. To solve the problems, the study proposes a dynamic hand gesture recognition framework on end devices. The authors provide a method (i.e., ModelOps) to deploy the trained model on end devices, by building an edge computing architecture using Kubernetes. The research provides developers with a real-time gesture recognition component. The experimental results show that the framework is suitable on end devices.
Article Preview
Top

1. Introduction

With the rapid development and popularization of computers and information technology, people can use their end devices (i.e., modern smartphones and Raspberry Pi) nearly anywhere, resulting in considerable research being devoted to the development of new applications for these ubiquitous end devices. Although these new applications provide significant benefits to users, their human–machine interfaces are still keyboards, mouses, or touch screens (Wang et al., 2017). Hand gesture recognition (Kim & Toomajian, 2016) can provide users with a more lively, natural, and convenient human–machine interface to operate and invoke applications on end devices. Also, hand gesture recognition can be used in human-robot interaction to create user interfaces that are natural to use and easy to learn. However, locating the hands and segmenting them from the background in an image sequence is a problem for hand gesture recognition.

In recent years, many studies (Costante et al., 2014; Dhingra & Kunz, 2019; Kim & Toomajian, 2016; Nanni et al., 2017; Shin & Sung, 2016; Žemgulys et al., 2018; ZOU et al., 2018) have used hand gesture recognition models for human–machine interface applications. These models are largely based on handcrafted features and feature extraction through deep learning. These models can be divided into static and dynamic gesture recognition. Static gesture recognition methods consider spatial features of hands, whereas dynamic gesture recognition methods extract not only spatial features but also temporal features.

In contrast to models based on handcrafted features, models (Costante et al., 2014; Dhingra & Kunz, 2019; Shin & Sung, 2016) based on deep learning perform well in automatic feature learning from image frames. Feature deep learning provides new insights into gesture recognition, and many researchers have attempted to use deep learning methods to extract gesture features from RGB, depth, and skeleton data. In (Shin & Sung, 2016), a dynamic hand gesture recognition technique was developed using a recurrent neural network (RNN) algorithm. In (Costante et al., 2014), deep CNNs and random forest (RF) algorithms were compared, and the results indicated that CNN slightly outperformed RF with sufficient data and achieved significantly better accuracy than other methods. A deep learning CNN can learn hand gesture features from single-mode data or multimodal fusion data. As the appearance and optical flow sequences are relatively easy to obtain, most deep learning methods adopt these two as their input, with few depth-based techniques.

Complete Article List

Search this Journal:
Reset
Volume 15: 1 Issue (2024)
Volume 14: 1 Issue (2023)
Volume 13: 8 Issues (2022): 7 Released, 1 Forthcoming
Volume 12: 4 Issues (2021)
Volume 11: 4 Issues (2020)
Volume 10: 4 Issues (2019)
Volume 9: 4 Issues (2018)
Volume 8: 4 Issues (2017)
Volume 7: 4 Issues (2016)
Volume 6: 4 Issues (2015)
Volume 5: 4 Issues (2014)
Volume 4: 4 Issues (2013)
Volume 3: 4 Issues (2012)
Volume 2: 4 Issues (2011)
Volume 1: 4 Issues (2010)
View Complete Journal Contents Listing