Four-class emotion classification in virtual reality using pupillometry

Lim Jia Zheng and James Mountstephens and Jason Teo (2020) Four-class emotion classification in virtual reality using pupillometry. Journal of Big Data, 7 (43). pp. 1-9.

[img]
Preview
Text
Four.pdf

Download (49kB) | Preview
[img] Text
s40537-020-00322-9.pdf
Restricted to Registered users only

Download (974kB)

Abstract

Background: Emotion classifcation remains a challenging problem in afective computing. The large majority of emotion classifcation studies rely on electroencephalography (EEG) and/or electrocardiography (ECG) signals and only classifes the emotions into two or three classes. Moreover, the stimuli used in most emotion classifcation studies utilize either music or visual stimuli that are presented through conventional displays such as computer display screens or television screens. This study reports on a novel approach to recognizing emotions using pupillometry alone in the form of pupil diameter data to classify emotions into four distinct classes according to Russell’s Circumplex Model of Emotions, utilizing emotional stimuli that are presented in a virtual reality (VR) environment. The stimuli used in this experiment are 360° videos presented using a VR headset. Using an eye-tracker, pupil diameter is acquired as the sole classifcation feature. Three classifers were used for the emotion classifcation which are Support Vector Machine (SVM), k-Nearest Neighbor (KNN), and Random Forest (RF). Findings: SVM achieved the best performance for the four-class intra-subject classifcation task at an average of 57.05% accuracy, which is more than twice the accuracy of a random classifer. Although the accuracy can still be signifcantly improved, this study reports on the frst systematic study on the use of eye-tracking data alone without any other supplementary sensor modalities to perform human emotion classifcation and demonstrates that even with a single feature of pupil diameter alone, emotions could be classifed into four distinct classes to a certain level of accuracy. Moreover, the best performance for recognizing a particular class was 70.83%, which was achieved by the KNN classifer for Quadrant 3 emotions. Conclusion: This study presents the frst systematic investigation on the use of pupillometry as the sole feature to classify emotions into four distinct classes using VR stimuli. The ability to conduct emotion classifcation using pupil data alone represents a promising new approach to afective computing as new applications could be developed using readily-available webcams on laptops and other mobile devices that are equipped with cameras without the need for specialized and costly equipment such as EEG and/or ECG as the sensor modality.

Item Type: Article
Keyword: Emotion classifcation, Eye-tracking, Pupil diameter, Virtual reality, Machinelearning
Subjects: ?? QA75 ??
Department: FACULTY > Faculty of Computing and Informatics
Depositing User: SITI AZIZAH BINTI IDRIS -
Date Deposited: 26 Oct 2020 20:29
Last Modified: 08 Apr 2021 22:36
URI: https://eprints.ums.edu.my/id/eprint/26209

Actions (login required)

View Item View Item