VR Presentation Training System Using Machine Learning Techniques for Automatic Evaluation

VR Presentation Training System Using Machine Learning Techniques for Automatic Evaluation

Yuto Yokoyama, Katashi Nagao
Copyright: © 2021 |Pages: 23
DOI: 10.4018/IJVAR.290044
OnDemand:
(Individual Articles)
Available
$37.50
No Current Special Offers
TOTAL SAVINGS: $37.50

Abstract

In this paper, the authors build an immersive training space using building-scale VR, a technology that makes a virtual space based on an entire building existing in the real world. The space is used for presentations, allowing students to self-train. The results of a presentation are automatically evaluated by using machine learning or the like and fed back to the user. In this space, users can meet their past selves (more accurately, their avatars), so they can objectively observe their presentations and recognize weak points. The authors developed a mechanism for recording and reproducing activities in virtual space in detail and a mechanism for applying machine learning to activity records. With these mechanisms, a system for recording, reproducing, and automatically evaluating presentations was developed.
Article Preview
Top

1. Introduction

The current use of VR is biased towards entertainment such as games, and much hardware and software development is for that purpose (Grubert et al., 2018; Gugenheimer et al., 2017; Huang et al., 2017). However, 6 degrees of freedom and controllers for interacting with the virtual world are also suitable for training, learning, and meetings, and we think that usage in these directions should be increased. In fact, the medical and athletic fields use VR simulation systems for surgical training and rehabilitation (Hamzeheinejad et al., 2018; Zahedi et al., 2017). In addition, research has been conducted on visualizing the movement of players in VR (Kaplan et al., 2018) and on improving the efficiency of VR training (Loreto et al., 2018; Isogawa et al., 2018).

Similarly, companies and universities are introducing VR conferences and VR classes, and their effectiveness has been verified (Clifford et al., 2018; Slavova et al., 2018; Volonte et al., 2018; Wienrich et al., 2018). In addition to being in a VR space remotely, VR-specific learning methods have been devised that display interactive 3D models or use 360-degree content for effective learning (Hussein et al., 2011).

However, VR usage related to meetings and classes mainly involves participation from remote locations with only a sense of reality, such as presence communication. Therefore, the VR feature of tracking the movement of parts of the body or the whole body is not faithfully taken advantage of.

Nonverbal communication channels such as gestures are also important in VR, and they are known to be particularly effective in presentations (Hostetter, 2011). Therefore, the body movements of VR users should be considered.

In an actual meeting, recording mainly involves video and audio. For this reason, extra work is required to obtain knowledge from these records, such as estimating the body movement of a speaker from images. However, in VR, the user always wears a tracking device, so data can be acquired more easily. In other words, a conference using VR is more likely to uncover the characteristics and customs of the participants than an in-person conference would.

Several systems for evaluating real-world presentations already exist (Kurihara et al., 2007). However, a methodology for improving the problems discovered and the specific solutions to the problems has not been established. The threshold of the evaluation indices used in the systems is also determined by intuition. Furthermore, gestures are not evaluated much during presentations.

In this paper, we propose a VR presentation training system that automatically evaluates and feeds back a user’s own presentations.

In this system, a presentation venue (conference hall) similar to one in reality is reproduced in VR, the presenter and audience are expressed as VR avatars, and the presentation is recorded in detail and evaluated automatically in real time. Speech/gaze are evaluated by Kurihara et al. (2007) and Wolf In Motion Ltd. (2016) in the same manner, and we also evaluate gestures in the same way. To evaluate gestures, we record the movements of presenters during presentations made in VR space and perform subjective evaluations with multiple subjects.

We perform machine learning with this data and develop an automatic presentation-evaluation and feedback system. One of the functions of the VR presentation training system is a VR playback function that browses past presentations in VR and encourages users to reflect on their own presentation performance. However, at present, there is no appropriate means of recording and analyzing past presentations and motions.

In this research, we developed VRec, a library that records activities in VR in detail, and VR2ML, a library that analyzes recorded data using machine learning (Yokoyama & Nagao, 2020), as a supplemental system for the VR presentation training system.

Complete Article List

Search this Journal:
Reset
Volume 7: 1 Issue (2024): Forthcoming, Available for Pre-Order
Volume 6: 2 Issues (2022): 1 Released, 1 Forthcoming
Volume 5: 2 Issues (2021)
Volume 4: 2 Issues (2020)
Volume 3: 2 Issues (2019)
Volume 2: 2 Issues (2018)
Volume 1: 2 Issues (2017)
View Complete Journal Contents Listing