Interactive Design of 3D Dynamic Gesture Based on SVM-LSTM Model

Interactive Design of 3D Dynamic Gesture Based on SVM-LSTM Model

Tao Wang, Xiaolong Cai, Liping Wang, Haoye Tian
Copyright: © 2018 |Pages: 15
DOI: 10.4018/IJMHCI.2018070104
OnDemand:
(Individual Articles)
Available
$37.50
No Current Special Offers
TOTAL SAVINGS: $37.50

Abstract

Visual hand gesture interaction is one of the main ways of human-computer interaction, and provides users more interactive degrees of freedom and more realistic interactive experience. Authors present a hybrid model based on SVM-LSTM, and design a three-dimensional dynamic gesture interaction system. The system uses Leap Motion to capture gesture information, combined with SVM powerful static gesture classification ability and LSTM powerful variable-length time series gesture processing ability, enabling real-time recognition of user gestures. The gesture interaction method can automatically define the start and end of gestures, recognition accuracy reached 96.4%, greatly reducing the cost of learning. Experiments have shown that the gesture interaction method proposed by authors is effective. In the simulated mobile environment, the average gesture prediction only takes 0.15 seconds, and ordinary users can quickly grasp this method.
Article Preview
Top

Introduction

With the development of human-computer interaction technology, a variety of new means of interaction are emerging, so that human-computer interaction is moving towards a more natural, efficient and intelligent direction. Visual hand gesture interaction is one of the main ways of human-computer interaction. Compared with the popular keyboard, mouse and interactive physical touch screen, visual gesture interaction can help users get rid of those constraints, and provide users greater interactive space, more interactive degrees of freedom and more realistic interactive experience. Therefore, the gesture interactive technology has attracted increasing attention around the world, and quickly became a hot research field of human-computer interaction. It has been widely used in virtual / augmented reality, pervasive computing, smart spaces, computer-based interactive games and other fields, such as Leap Motion, Microsoft Kinect and Intel RealSence have quietly stepped into the lives of consumers (Si & Hou, 2016; Wang et al., 2016; Chao et al., 2017).

At the same time, the rapid development of augmented reality is more realistic for the interaction of gestures. At present, augmented reality mainly takes the mobile phone and the head-mounted display as a hardware carrier. The head-mounted display has a stronger immersion than the mobile phone, and the overall experience is even better. Different from the mature touch screen interaction mode on the mobile phone, the head-mounted display will use the natural gesture interaction as the main interaction mode. Because, as compared with conventional virtual reality head-mounted display handle operation, the use of an augmented reality scene is more diverse, more portable equipment requirements. After all, people can accept playing virtual reality games with handle in living room, but do not want to use the handle when operating an augmented reality device outdoors. At the same time, 3D dynamic gestures are one of the most natural gestures of human interaction. Obviously, users want to spend a lower learning costs to use augmented reality device with the natural gestures (Yaoyuneyong, Foster, & Flynn, 2014; Waghmare Amit, Sonawane Kunal, Chavan Puja, & Kadu Nanasaheb, 2014).

Traditional gesture interaction techniques are often based on a statistical learning method or neural network technology, most of these methods have some problems such as poor recognition accuracy, high learning cost and so on. Li et al. (2012) uses Three-point Alignment Algorithm in Kinect to recognize planar gestures with accuracy 84%; Hasan et al. (2014) presents a method for hand gesture recognition using neural networks, which obtain 86.3% accuracy; Dardas et al. (2011) presents a real time system, which includes detecting and tracking bare hand in cluttered background using skin detection and hand postures contours comparison algorithm after face subtraction, and recognizing hand gestures using Principle Components Analysis (PCA) and obtain 90% accuracy; Trigueiros et al. (2013) tries a few classifiers, the highest accuracy classifier is neural networks (accuracy 91%); Lai et al. (2014) presents a method based on convex defect character points of the hand contour, recognizing planar number gestures in real-time with 95.1% accuracy, but planar digital gestures are not natural human-computer interaction gestures; Wang et al. (2013) uses PCA+LDA algorithm in 3d laser scanner to obtain 96% accuracy, but cannot inference in real time.

Complete Article List

Search this Journal:
Reset
Volume 16: 1 Issue (2024): Forthcoming, Available for Pre-Order
Volume 15: 1 Issue (2023)
Volume 14: 4 Issues (2022): 1 Released, 3 Forthcoming
Volume 13: 1 Issue (2021)
Volume 12: 3 Issues (2020)
Volume 11: 4 Issues (2019)
Volume 10: 4 Issues (2018)
Volume 9: 4 Issues (2017)
Volume 8: 4 Issues (2016)
Volume 7: 4 Issues (2015)
Volume 6: 4 Issues (2014)
Volume 5: 4 Issues (2013)
Volume 4: 4 Issues (2012)
Volume 3: 4 Issues (2011)
Volume 2: 4 Issues (2010)
Volume 1: 4 Issues (2009)
View Complete Journal Contents Listing