Predicting user visual attention in virtual reality with a deep learning model

12Citations
Citations of this article
40Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Recent studies show that user’s visual attention during virtual reality museum navigation can be effectively estimated with deep learning models. However, these models rely on large-scale datasets that usually are of high structure complexity and context specific, which is challenging for nonspecialist researchers and designers. Therefore, we present the deep learning model, ALRF, to generalise on real-time user visual attention prediction in virtual reality context. The model combines two parallel deep learning streams to process the compact dataset of temporal–spatial salient features of user’s eye movements and virtual object coordinates. The prediction accuracy outperformed the state-of-the-art deep learning models by reaching record high 91.03%. Importantly, with quick parametric tuning, the model showed flexible applicability across different environments of the virtual reality museum and outdoor scenes. Implications for how the proposed model may be implemented as a generalising tool for adaptive virtual reality application design and evaluation are discussed.

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Li, X., Shan, Y., Chen, W., Wu, Y., Hansen, P., & Perrault, S. (2021). Predicting user visual attention in virtual reality with a deep learning model. Virtual Reality, 25(4), 1123–1136. https://doi.org/10.1007/s10055-021-00512-7

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 10

59%

Lecturer / Post doc 3

18%

Researcher 3

18%

Professor / Associate Prof. 1

6%

Readers' Discipline

Tooltip

Computer Science 6

38%

Business, Management and Accounting 4

25%

Design 3

19%

Engineering 3

19%

Save time finding and organizing research with Mendeley

Sign up for free