[go: nahoru, domu]

Skip to content

r-zemblys/EM-event-detection-evaluation

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

12 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Evaluating Eye Movement Event Detection: A Review of the State of the Art

Dissemination

@article{startsev2022evaluating,
  title={Evaluating Eye Movement Event Detection: A Review of the State of the Art},
  author={Startsev, Mikhail and Zemblys, Raimondas},
  journal={Behavior Research Methods},
  pages={1--62},
  year={2022},
  publisher={Springer}
}

Full paper is available via https://link.springer.com/article/10.3758/s13428-021-01763-7.

ETRA2022 tutorial slides are available via https://emdetectionmetrics.page.link/etra2022. IPython Notebook demo available here: demo/ETRA2022 Demo.ipynb.

ETRA2024 tutorial slides are available here. IPython Notebook demo available here: demo/ETRA2024 Demo.ipynb

Run

python run_calc.py

Replicating paper results

  1. Download Hollywood2EM (Agtzidis et al., 2020) dataset from https://gin.g-node.org/ioannis.agtzidis/hollywood2_em

  2. Run the following to convert dataset to the required format:

python misc/scripts/run_data_parse.py -root DATASET_ROOT -dataset hollywood2em --coder expert 
python misc/scripts/run_data_parse.py -root DATASET_ROOT -dataset hollywood2em --coder alg 

DATASET_ROOT is the directory where dataset was downloaded

  1. Run evaluation script:
python run_calc.py  -job assets/job_hollywood2.json

The result file will be save to ./results/job_hollywood2.csv. This can take around 3 hours or more, depending on the computer

  1. Run data analysis script:
python misc/scripts/analyse_results.py

Resulting plots will be saved to ./results/job_hollywood2/result-plots

To get matcher example plots run:

python run_calc.py -job assets/job_plot.json

Resulting plots will be saved to ./results/match-plots

List of publicly available annotated eye movement datasets

List of publicly available annotated datasets to illustrate the variety of readily available material for algorithm development and evaluation. Duration reflects the amount of unique eye-tracking data (uniqueness judged based on file names); duration in parentheses -- the total amount of available annotated data (including undefined samples and taking into account several available annotations for a single recording). Sample distributions do not list proportion of undefined samples and samples annotated as noise, blinks and similar. Note that datasets might have different definitions of fixations, saccades, and other events.

Dataset Duration Set-up Sampling frequency Eye-tracker Sample distribution
Lund2013
(Andersson et al., 2017)
14.9 min
(18.6 min)
Screen-based, pictures, moving dots and video clips 500 Hz SMI Hi-Speed 1250 46.49% Fixation
5.88% Saccade
3.34% PSO
41.60% Pursuit
Notes: two expert annotators, fully manually annotated, partial annotation overlap. Includes data that was used in other papers. Download from https://github.com/richardandersson/EyeMovementDetectorEvaluation
IRF
(Zemblys et al., 2018)
8.1 min Screen-based, fixate-saccade task 1000 Hz EyeLink 1000Plus 86.77% Fixation
5.65% Saccade
3.00% PSO
Notes: one expert annotator, fully manually annotated. Six participants, data from a replication study. Download from https://github.com/r-zemblys/irf
MPIIEgoFixation
(Steil et al., 2018)
24.2 min Head-mounted, unscripted daily life activities 30 Hz Pupil Pro 74.19% Fixation
Notes: frame-by-frame annotations of one annotator. Download from https://www.mpi-inf.mpg.de/MPIIEgoFixation
humanFixationClassification
(Hooge et al., 2018)
5.9 min
(70.4 min)
Screen-based, pictures and search task 300 Hz Tobii TX300 71.82% Fixation
Notes: 12 expert annotators, fully manually annotated, all annotation data overlap. 10 adult free viewing and 60 infant search task trials. Download from https://github.com/dcnieho/humanFixationClassification
360EM
(Agtzidis et al., 2019)
32.9 min Head-mounted, naturalistic 360° videos 120 Hz FOVE Primary labels:
75.15% Fixation
10.44% Saccade
9.76% Pursuit

Secondary labels:
0.81% OKN
27.64% VOR
15.84% OKN+VOR
1.47% Head pursuit
Notes: two stage annotations of one expert annotator after training and discussion session. First stage (primary labels and optokinetic nystagmus - OKN - or nystagmus) uses pre-labelled saccades and does not account for the head motion. Second stage (vestibulo-ocular reflex - VOR, VOR + OKN, Head pursuit) uses labels from the previous stage that are re-examined in the context of the eye-head coordination. Ca. 3.5 h of eye- and head-tracking recordings, ca. 16% annotated. Download from https://gin.g-node.org/ioannis.agtzidis/360_em_dataset
GazeCom
(Startsev et al., 2019)
4.7 h
(14.1 h)
Screen-based, naturalistic video 250 Hz EyeLink II 73.96% Fixation
10.67% Saccade
9.83% Pursuit
Notes: manual annotations of one expert tie-breaking and adjusting labels of two novice annotators. Novice annotators (paid undergraduate students) used pre-labeled data and went through the data twice. Labels of novice annotators are available. Download from https://gin.g-node.org/ioannis.agtzidis/gazecom_annotations
Hollywood2EM
(Agtzidis et al., 2020)
2.15 h
(4.3 h)
Screen-based, movie clips 500 Hz SMI Hi-Speed 1250 59.46% Fixation
9.87% Saccade
26.54% Pursuit
Notes: manual annotations of pre-labeled data, two stage annotation (paid student followed by an expert coder). Labels of student annotator are available. Download from https://gin.g-node.org/ioannis.agtzidis/hollywood2_em
Gaze-in-wild
(Kothatri et al., 2020)
3.06 h
(4.15 h)
Head mounted, naturalistic tasks 300 Hz Pupil Labs + custom setup 12.50% Fixation
7.12% Saccade
2.65% Pursuit
26.72% VOR
Notes: independent annotations of five trained annotators, ca. half of the data is annotated. Naturalistic tasks: indoor navigation, ball catching, object search, tea making. Download from http://www.cis.rit.edu/~rsk3900/gaze-in-wild

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages