/eye_track_spatiotemporal

A lightweight spatio-temporal network for online eye tracking with event camera

Primary LanguagePython

A lightweight spatio-temporal network for online eye tracking with event camera

Instructions

Set up a virtual environment and install the required dependancies as follows:

conda create -n eye_tracking python==3.10
conda activate eye_tracking
pip install -r requirements.txt

To reproduce our final submission.csv to the Event-based Eye Tracking - AIS2024 CVPR Workshop, first download the data from the competition page and place the .zip file in the root of this repository then run the following to extract the data

unzip event-based-eye-tracking-ais2024.zip -d temp
mv temp/event_data/event_data event_data
rm event-based-eye-tracking-ais2024.zip
rm -r temp

Then simply run python generate_submission.py to perform inference with our model (no GPU needed), which uses our checkpoint in weights/submission.ckpt and config in submission_config.yaml to generate our submission.

We use hydra to mange our configs during development and ablation studies. The main entry point to a training run is train.py, which will launch a run with the config in config.yaml.

Contributors

Rudy Pei: yanrpei@gmail.com

Sasskia Brüers: s.bruers@gmail.com

Sébastien Crouzet: seb.crouzet@gmail.com

Douglas McLelland: dmclelland@brainchip.com