Gnacek, M., Quintero, L., Mavridou, I., Balaguer-Ballester, E., Kostoulas, T., Nduka, C. and Seiss, E., 2024. AVDOS-VR: Affective Video Database with Physiological Signals and Continuous Ratings Collected Remotely in VR. Scientific Data, 11, 132.
Full text available as:
|
PDF (OPEN ACCESS ARTICLE)
s41597-024-02953-6.pdf - Published Version Available under License Creative Commons Attribution. 2MB | |
Copyright to original material in this document is with the original owner(s). Access to this content through BURO is granted on condition that you use it only for research, scholarly or other non-commercial purposes. If you wish to use it for any other purposes, you must contact BU via BURO@bournemouth.ac.uk. Any third party copyright material in this document remains the property of its respective owner(s). BU grants no licence for further use of that third party material. |
Official URL: https://www.nature.com/sdata/
DOI: 10.1038/s41597-024-02953-6
Abstract
Investigating emotions relies on pre-validated stimuli to evaluate induced responses through subjective self-ratings and physiological changes. The creation of precise affect models necessitates extensive datasets. While datasets related to pictures, words, and sounds are abundant, those associated with videos are comparatively scarce. To overcome this challenge, we present the first virtual reality (VR) database with continuous self-ratings and physiological measures, including facial EMG. Videos were rated online using a head-mounted VR device (HMD) with attached emteqPRO mask and a cinema VR environment in remote home and laboratory settings with minimal setup requirements. This led to an affective video database with continuous valence and arousal self-rating measures and physiological responses (PPG, facial-EMG (7x), IMU). The AVDOS-VR database includes data from 37 participants who watched 30 randomly ordered videos (10 positive, neutral, and negative). Each 30-second video was assessed with two-minute relaxation between categories. Validation results suggest that remote data collection is ecologically valid, providing an effective strategy for future affective study designs. All data can be accessed via: www.gnacek.com/affective-video-database-online-study
Item Type: | Article |
---|---|
ISSN: | 2052-4463 |
Uncontrolled Keywords: | VR; Virtual Reality; facial-EMG; PPG; IMU; Affect Detection; Machine Learning for Affect Classification |
Group: | Faculty of Science & Technology |
ID Code: | 39384 |
Deposited By: | Symplectic RT2 |
Deposited On: | 11 Jan 2024 11:13 |
Last Modified: | 30 Jan 2024 08:42 |
Downloads
Downloads per month over past year
Repository Staff Only - |