Home > Research > Publications & Outputs > Real-time head-based deep-learning model for ga...

Electronic data

  • Accepted Manuscript

    Rights statement: © ACM, 2022. This is the author's version of the work. It is posted here by permission of ACM for your personal use. Not for redistribution. The definitive version was published in ETRA '22: 2022 Symposium on Eye Tracking Research and Applications https://dl.acm.org/doi/10.1145/3517031.3529642

    Accepted author manuscript, 1.52 MB, PDF document

    Available under license: CC BY-NC: Creative Commons Attribution-NonCommercial 4.0 International License

Links

Text available via DOI:

View graph of relations

Real-time head-based deep-learning model for gaze probability regions in collaborative VR

Research output: Contribution in Book/Report/Proceedings - With ISBN/ISSNConference contribution/Paperpeer-review

Published
Close
Publication date11/06/2022
Host publicationACM Symposium on Eye Tracking Research and Applications
Place of PublicationNew York
PublisherACM
Number of pages8
ISBN (electronic)9781450392525
<mark>Original language</mark>English
EventETRA '22: 2022 Symposium on Eye Tracking Research and Applications - Seattle Children’s Building Cure , Seattle, United States
Duration: 8/06/202211/06/2022
https://etra.acm.org/2022/

Symposium

SymposiumETRA '22: 2022 Symposium on Eye Tracking Research and Applications
Country/TerritoryUnited States
CitySeattle
Period8/06/2211/06/22
Internet address

Symposium

SymposiumETRA '22: 2022 Symposium on Eye Tracking Research and Applications
Country/TerritoryUnited States
CitySeattle
Period8/06/2211/06/22
Internet address

Abstract

Eye behaviour has gained much interest in the VR research community as an interaction input and support for collaboration. Researchers implemented gaze inference models when eye-tracking is missing by using head behavior and saliency. However, these solutions are resource-demanding and thus unfit for untethered devices, and their angle accuracy is around 7°, which can be a problem in high-density informative areas. To address this issue, we propose a lightweight deep learning model that generates the probability density function of the gaze as a percentile contour. This solution allows us to introduce a visual attention representation based on a region rather than a point and manage a trade-off between the ambiguity of a region and the error of a point. We tested our model in untethered devices with real-time performances; we evaluated its accuracy which outperforms our identified baselines (average fixation map and head direction).

Bibliographic note

© ACM, 2022. This is the author's version of the work. It is posted here by permission of ACM for your personal use. Not for redistribution. The definitive version was published in ETRA '22: 2022 Symposium on Eye Tracking Research and Applications https://dl.acm.org/doi/10.1145/3517031.3529642