Investigation of eye movement strategies often requires the measurement of gaze orientation without restraining the head. However, most commercial eye-trackers have low tolerance for head movements. Here we present a novel geometry-based method to estimate gaze orientation in space in unrestricted head conditions. The method combines the measurement of eye-in-head orientation-provided by a head-mounted video-based eye-tracker-and head-in-space position and orientation-provided by a motion capture system. The method does not rely on specific assumptions on the configuration of the eye-tracker camera with respect to the eye and uses a central projection to estimate the pupil position from the camera image, thus improving upon previously proposed geometry-based procedures. The geometrical parameters for the mapping between pupil image and gaze orientation are derived with a calibration procedure based on nonlinear constrained optimization. Additionally, the method includes a procedure to correct for possible slippages of the tracker helmet based on a geometrical representation of the pupil-to-gaze mapping. We tested and validated our method on seven subjects in the context of a one-handed catching experiment. We obtained accuracy better than 0.8° and precision better than 0.5° in the measurement of gaze orientation. Our method can be used with any video-based eye-tracking system to investigate eye movement strategies in a broad range of naturalistic experimental scenarios.

Cesqui, B., de Langenberg, R., Lacquaniti, F., D'Avella, A. (2013). A novel method for measuring gaze orientation in space in unrestrained head conditions. JOURNAL OF VISION, 13(8) [10.1167/13.8.28].

A novel method for measuring gaze orientation in space in unrestrained head conditions

LACQUANITI, FRANCESCO;
2013-07-31

Abstract

Investigation of eye movement strategies often requires the measurement of gaze orientation without restraining the head. However, most commercial eye-trackers have low tolerance for head movements. Here we present a novel geometry-based method to estimate gaze orientation in space in unrestricted head conditions. The method combines the measurement of eye-in-head orientation-provided by a head-mounted video-based eye-tracker-and head-in-space position and orientation-provided by a motion capture system. The method does not rely on specific assumptions on the configuration of the eye-tracker camera with respect to the eye and uses a central projection to estimate the pupil position from the camera image, thus improving upon previously proposed geometry-based procedures. The geometrical parameters for the mapping between pupil image and gaze orientation are derived with a calibration procedure based on nonlinear constrained optimization. Additionally, the method includes a procedure to correct for possible slippages of the tracker helmet based on a geometrical representation of the pupil-to-gaze mapping. We tested and validated our method on seven subjects in the context of a one-handed catching experiment. We obtained accuracy better than 0.8° and precision better than 0.5° in the measurement of gaze orientation. Our method can be used with any video-based eye-tracking system to investigate eye movement strategies in a broad range of naturalistic experimental scenarios.
31-lug-2013
Pubblicato
Rilevanza internazionale
Articolo
Esperti anonimi
Settore BIO/09 - FISIOLOGIA
English
Con Impact Factor ISI
gaze orientation in space; nonlinear optimization; calibration; oculography; drift correction
Cesqui, B., de Langenberg, R., Lacquaniti, F., D'Avella, A. (2013). A novel method for measuring gaze orientation in space in unrestrained head conditions. JOURNAL OF VISION, 13(8) [10.1167/13.8.28].
Cesqui, B; de Langenberg, R; Lacquaniti, F; D'Avella, A
Articolo su rivista
File in questo prodotto:
File Dimensione Formato  
CESQUI_A novel method _2013.pdf

accesso aperto

Descrizione: Articolo principale
Dimensione 2.15 MB
Formato Adobe PDF
2.15 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/2108/79565
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 18
  • ???jsp.display-item.citation.isi??? 13
social impact