skip to main content
10.1145/3314111.3319822acmconferencesArticle/Chapter ViewAbstractPublication PagesetraConference Proceedingsconference-collections
research-article

Monocular gaze depth estimation using the vestibulo-ocular reflex

Published: 25 June 2019 Publication History

Abstract

Gaze depth estimation presents a challenge for eye tracking in 3D. This work investigates a novel approach to the problem based on eye movement mediated by the vestibulo-ocular reflex (VOR). VOR stabilises gaze on a target during head movement, with eye movement in the opposite direction, and the VOR gain increases the closer the fixated target is to the viewer. We present a theoretical analysis of the relationship between VOR gain and depth which we investigate with empirical data collected in a user study (N=10). We show that VOR gain can be captured using pupil centres, and propose and evaluate a practical method for gaze depth estimation based on a generic function of VOR gain and two-point depth calibration. The results show that VOR gain is comparable with vergence in capturing depth while only requiring one eye, and provide insight into open challenges in harnessing VOR gain as a robust measure.

Supplementary Material

ZIP File (a20-mardenbegi.zip)
Supplemental files.

References

[1]
Florian Alt, Stefan Schneegass, Jonas Auda, Rufat Rzayev, and Nora Broy. 2014a. Using Eye-tracking to Support Interaction with Layered 3D Interfaces on Stereoscopic Displays. In Proceedings of the 19th International Conference on Intelligent User Interfaces (IUI '14). ACM, New York, NY, USA, 267--272.
[2]
Florian Alt, Stefan Schneegass, Jonas Auda, Rufat Rzayev, and Nora Broy. 2014b. Using Eye-tracking to Support Interaction with Layered 3D Interfaces on Stereoscopic Displays. In Proceedings of the 19th International Conference on Intelligent User Interfaces (IUI '14). ACM, New York, NY, USA, 267--272.
[3]
Dora E Angelaki. 2004. Eyes on target: what neurons must do for the vestibuloocular reflex during linear motion. Journal of neurophysiology 92, 1 (2004), 20--35.
[4]
B. Biguer and C. Prablanc. 1981. Modulation of the vestibulo-ocular reflex in eye-head orientation as a function of target distance in man. Progress in Oculomotor Research (1981). https://ci.nii.ac.jp/naid/10008955589/en/
[5]
Gilles Clément and Fernanda Maciel. 2004. Adjustment of the vestibulo-ocular reflex gain as a function of perceived target distance in humans. Neuroscience letters 366, 2 (2004), 115--119.
[6]
Han Collewijn and Jeroen BJ Smeets. 2000. Early components of the human vestibulo-ocular response to head rotation: latency and gain. Journal of Neurophysiology 84, 1 (2000), 376--389.
[7]
Nathan Cournia, John D Smith, and Andrew T Duchowski. 2003. Gaze-vs. hand-based pointing in virtual environments. In CHI'03 extended abstracts on Human factors in computing systems. ACM, 772--773.
[8]
Brian C. Daugherty, Andrew T. Duchowski, Donald H. House, and Celambarasan Ramasamy. 2010. Measuring Vergence over Stereoscopic Video with a Remote Eye Tracker. In Proceedings of the 2010 Symposium on Eye-Tracking Research and Applications (ETRA '10). ACM, New York, NY, USA, 97--100.
[9]
S. Deng, J. Chang, S. Hu, and J. J. Zhang. 2017. Gaze Modulated Disambiguation Technique for Gesture Control in 3D Virtual Objects Selection. In 2017 3rd IEEE International Conference on Cybernetics (CYBCONF). 1--8.
[10]
Andrew T. Duchowski, Donald H. House, Jordan Gestring, Robert Congdon, Lech Świrski, Neil A. Dodgson, Krzysztof Krejtz, and Izabela Krejtz. 2014. Comparing Estimated Gaze Depth in Virtual and Physical Environments. In Proceedings of the Symposium on Eye Tracking Research and Applications (ETRA '14). ACM, New York, NY, USA, 103--110.
[11]
Andrew T. Duchowski, Eric Medlin, Anand Gramopadhye, Brian Melloy, and Santosh Nair. 2001. Binocular Eye Tracking in VR for Visual Inspection Training. In Proceedings of the ACM Symposium on Virtual Reality Software and Technology (VRST '01). ACM, New York, NY, USA, 1--8.
[12]
Andrew T. Duchowski, Brandon Pelfrey, Donald H. House, and Rui Wang. 2011. Measuring Gaze Depth with an Eye Tracker During Stereoscopic Display. In Proceedings of the ACM SIGGRAPH Symposium on Applied Perception in Graphics and Visualization (APGV '11). ACM, New York, NY, USA, 15--22.
[13]
Peter A Gorry. 1990. General least-squares smoothing and differentiation by the convolution (Savitzky-Golay) method. Analytical Chemistry 62, 6 (1990), 570--573.
[14]
Esteban Gutierrez Mlot, Hamed Bahmani, Siegfried Wahl, and Enkelejda Kasneci. 2016. 3D Gaze Estimation Using Eye Vergence. In Proceedings of the International Joint Conference on Biomedical Engineering Systems and Technologies (BIOSTEC 2016). SCITEPRESS - Science and Technology Publications, Lda, Portugal, 125--131.
[15]
C. Hennessey<sup>*</sup> and P. Lawrence. 2009. Noncontact Binocular Eye-Gaze Tracking for Point-of-Gaze Estimation in Three Dimensions. IEEE Transactions on Biomedical Engineering 56, 3 (March 2009), 790--799.
[16]
Trevor Hine and Frank Thorn. 1987. Compensatory eye movements during active head rotation for near targets: effects of imagination, rapid head oscillation and vergence. Vision research 27, 9 (1987), 1639--1657.
[17]
J. Ki and Y. Kwon. 2008. 3D Gaze Estimation and Interaction. In 2008 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video. 373--376.
[18]
Yong-Moo Kwon, Kyeong-Won Jeon, Jeongseok Ki, Qonita M Shahab, Sangwoo Jo, and Sung-Kyu Kim. 2006. 3D Gaze Estimation and Interaction to Stereo Dispaly. IJVR 5, 3 (2006), 41--45.
[19]
Radosław Mantiuk, Bartosz Bazyluk, and Anna Tomaszewska. 2011. Gaze-Dependent Depth-of-Field Effect Rendering in Virtual Environments. In Serious Games Development and Applications, Minhua Ma, Manuel Fradinho Oliveira, and João Madeiras Pereira (Eds.). Springer Berlin Heidelberg, Berlin, Heidelberg, 1--12.
[20]
Diako Mardanbegi, Dan Witzner Hansen, and Thomas Pederson. 2012. Eye-based Head Gestures. In Proceedings of the Symposium on Eye Tracking Research and Applications (ETRA '12). ACM, New York, NY, USA, 139--146.
[21]
Diako Mardanbegi, Tobias Langlotz, and Hans Gellersen. 2019. Resolving Target Ambiguity in 3D Gaze Interaction through VOR Depth Estimation. In CHI'19 Proceedings on Human Factors in Computing Systems.
[22]
Olivier Mercier, Yusufu Sulai, Kevin Mackenzie, Marina Zannoli, James Hillis, Derek Nowrouzezahrai, and Douglas Lanman. 2017. Fast Gaze-contingent Optimal Decompositions for Multifocal Displays. ACM Trans. Graph. 36, 6, Article 237 (Nov. 2017), 15 pages.
[23]
Susan M. Munn and Jeff B. Pelz. 2008. 3D Point-of-regard, Position and Head Orientation from a Portable Monocular Video-based Eye Tracker. In Proceedings of the 2008 Symposium on Eye Tracking Research & Applications (ETRA '08). ACM, New York, NY, USA, 181--188.
[24]
Tomi Nukarinen, Jari Kangas, Oleg Špakov, Poika Isokoski, Deepak Akkil, Jussi Rantala, and Roope Raisamo. 2016. Evaluation of HeadTurn: An Interaction Technique Using the Gaze and Head Turns. In Proceedings of the 9th Nordic Conference on Human-Computer Interaction (NordiCHI '16). ACM, New York, NY, USA, Article 43, 8 pages.
[25]
Jason Orlosky, Takumi Toyama, Daniel Sonntag, and Kiyoshi Kiyokawa. 2016. The Role of Focus in Advanced Visual Interfaces. KI - Künstliche Intelligenz 30, 3 (01 Oct 2016), 301--310.
[26]
Gary D Paige. 1989. The influence of target distance on eye movement responses during vertical linear motion. Experimental Brain Research 77, 3 (1989), 585--593.
[27]
Thies Pfeiffer, Marc Erich Latoschik, and Ipke Wachsmuth. 2008. Evaluation of binocular eye trackers and algorithms for 3D gaze interaction in virtual reality environments. JVRB-Journal of Virtual Reality and Broadcasting 5, 16 (2008).
[28]
A Poston. 2000. Static adult human physical characteristics of the adult head. Department of Defense Human Factors Engineering Technical Advisory Group (DOD-HDBK-743A) pp 72 (2000), 75.
[29]
Mina Ranjbaran and Henrietta L Galiana. 2012. The horizontal angular vestibulo-ocular reflex: a non-linear mechanism for context-dependent responses. In 2012 Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE, 3866--3869.
[30]
Gerald FÃČÅŠtterer Norbert Leister Stephan Reichelt, Ralf HÃČâĆňussler. 2010. Depth cues in human visual perception and their realization in 3D displays. (2010).
[31]
Vildan Tanriverdi and Robert JK Jacob. 2000. Interacting with eye movements in virtual environments. In Proceedings of the SIGCHI conference on Human Factors in Computing Systems. ACM, 265--272.
[32]
E Viirre, D Tweed, K Milner, and T Vilis. 1986. A reexamination of the gain of the vestibuloocular reflex. Journal of Neurophysiology 56, 2 (1986), 439--450.
[33]
Rui I. Wang, Brandon Pelfrey, Andrew T. Duchowski, and Donald H. House. 2014. Online 3D Gaze Localization on Stereoscopic Displays. ACM Trans. Appl. Percept. 11, 1, Article 3 (April 2014), 21 pages.
[34]
Martin Weier, Thorsten Roth, André Hinkenjann, and Philipp Slusallek. 2018. Predicting the Gaze Depth in Head-mounted Displays Using Multiple Feature Regression. In Proceedings of the 2018 ACM Symposium on Eye Tracking Research & Applications (ETRA '18). ACM, New York, NY, USA, Article 19, 9 pages.

Cited By

View all
  • (2024)WatchCap: Improving Scanning Efficiency in People with Low Vision through Compensatory Head Movement StimulationProceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies10.1145/36595928:2(1-32)Online publication date: 15-May-2024
  • (2023)GazeRayCursor: Facilitating Virtual Reality Target Selection by Blending Gaze and Controller RaycastingProceedings of the 29th ACM Symposium on Virtual Reality Software and Technology10.1145/3611659.3615693(1-11)Online publication date: 9-Oct-2023
  • (2023)Vergence Matching: Inferring Attention to Objects in 3D Environments for Gaze-Assisted SelectionProceedings of the 2023 CHI Conference on Human Factors in Computing Systems10.1145/3544548.3580685(1-15)Online publication date: 19-Apr-2023
  • Show More Cited By

Index Terms

  1. Monocular gaze depth estimation using the vestibulo-ocular reflex

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    ETRA '19: Proceedings of the 11th ACM Symposium on Eye Tracking Research & Applications
    June 2019
    623 pages
    ISBN:9781450367097
    DOI:10.1145/3314111
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 25 June 2019

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. 3D gaze estimation
    2. VOR
    3. eye movement
    4. eye tracking
    5. fixation depth
    6. gaze depth estimation

    Qualifiers

    • Research-article

    Funding Sources

    • EPSRC

    Conference

    ETRA '19

    Acceptance Rates

    Overall Acceptance Rate 69 of 137 submissions, 50%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)47
    • Downloads (Last 6 weeks)2
    Reflects downloads up to 25 Aug 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)WatchCap: Improving Scanning Efficiency in People with Low Vision through Compensatory Head Movement StimulationProceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies10.1145/36595928:2(1-32)Online publication date: 15-May-2024
    • (2023)GazeRayCursor: Facilitating Virtual Reality Target Selection by Blending Gaze and Controller RaycastingProceedings of the 29th ACM Symposium on Virtual Reality Software and Technology10.1145/3611659.3615693(1-11)Online publication date: 9-Oct-2023
    • (2023)Vergence Matching: Inferring Attention to Objects in 3D Environments for Gaze-Assisted SelectionProceedings of the 2023 CHI Conference on Human Factors in Computing Systems10.1145/3544548.3580685(1-15)Online publication date: 19-Apr-2023
    • (2022)DEEP: 3D Gaze Pointing in Virtual Reality Leveraging Eyelid MovementProceedings of the 35th Annual ACM Symposium on User Interface Software and Technology10.1145/3526113.3545673(1-14)Online publication date: 29-Oct-2022
    • (2022)Gaze-Vergence-Controlled See-Through Vision in Augmented RealityIEEE Transactions on Visualization and Computer Graphics10.1109/TVCG.2022.320311028:11(3843-3853)Online publication date: Nov-2022
    • (2020)Outline Pursuits: Gaze-assisted Selection of Occluded Objects in Virtual RealityProceedings of the 2020 CHI Conference on Human Factors in Computing Systems10.1145/3313831.3376438(1-13)Online publication date: 21-Apr-2020
    • (2019)Eye, Head and Torso Coordination During Gaze Shifts in Virtual RealityACM Transactions on Computer-Human Interaction10.1145/336121827:1(1-40)Online publication date: 14-Dec-2019

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media

    -