Skip to main content
Log in

Predicting user visual attention in virtual reality with a deep learning model

  • Original Article
  • Published:
Virtual Reality Aims and scope Submit manuscript

A Correction to this article was published on 28 April 2021

This article has been updated

Abstract

Recent studies show that user’s visual attention during virtual reality museum navigation can be effectively estimated with deep learning models. However, these models rely on large-scale datasets that usually are of high structure complexity and context specific, which is challenging for nonspecialist researchers and designers. Therefore, we present the deep learning model, ALRF, to generalise on real-time user visual attention prediction in virtual reality context. The model combines two parallel deep learning streams to process the compact dataset of temporal–spatial salient features of user’s eye movements and virtual object coordinates. The prediction accuracy outperformed the state-of-the-art deep learning models by reaching record high 91.03%. Importantly, with quick parametric tuning, the model showed flexible applicability across different environments of the virtual reality museum and outdoor scenes. Implications for how the proposed model may be implemented as a generalising tool for adaptive virtual reality application design and evaluation are discussed.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9

Similar content being viewed by others

Availability of data and material

Derived data supporting the findings of this study are available from the corresponding author on request.

Code availability

Software applications used in the study are based on public open sources, and the code being used in this study is available from the corresponding author on request.

Change history

References

Download references

Acknowledgements

The work is supported by Natural Science Foundation of China (61802341) and ZJU-SUTD IDEA programme (IDEA006).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xiangdong Li.

Ethics declarations

Conflict of interest

The authors report no conflicts of interest.

Ethical approval

This study was approved by the university human research ethics committee, and all procedures performed in studies involving human participants were in accordance with the ethical standards of the institutional and/or national research committee and with the 1964 Helsinki declaration and its later amendments or comparable ethical standards.

Informed consent

Informed consent was obtained from all individual participants included in the study.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

The Original article has been corrected: The fifth author name was corrected as Praben Hansen.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Li, X., Shan, Y., Chen, W. et al. Predicting user visual attention in virtual reality with a deep learning model. Virtual Reality 25, 1123–1136 (2021). https://doi.org/10.1007/s10055-021-00512-7

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10055-021-00512-7

Keywords

Navigation