Abstract
Virtual reality (VR), particularly through 3D presentations, significantly boosts user engagement and task efficiency in fields such as gaming, education, and healthcare, offering more immersive and interactive experiences than traditional 2D formats. This study investigates EEG classification in response to 2D and 3D VR stimuli to deepen our understanding of the neural mechanisms driving VR interactions, with implications for brain-computer interfaces (BCIs). We introduce STA-EEGNet, an innovative model that enhances EEGNet by incorporating spatial-temporal attention (STA), improving EEG signal classification from VR environments. A one-way analysis of variance (ANOVA) was utilized to optimize channel selection, enhancing model accuracy. Comparative experiments showed that STA-EEGNet surpassed traditional EEGNet, achieving a peak accuracy of 99.78 % with channel selection. These findings highlight the benefits of spatial-temporal attention and optimal channel selection in classifying VR-evoked EEG data. This study underscores the importance of integrating spatial-temporal attention with compact convolutional neural networks like EEGNet, not only improving EEG signal classification but also advancing neural decoding and optimizing BCI applications.