Datasets
This page provides the dataset-related information of 2D gaze estimation
.
In this page,
- we introduce the data pre-processing of each datasets.
- we provide the code of data pre-processing.
- we summarize the links of datasets.
We invite you to use our data processing code, and add your result into the benchmark.
Note that, you should first download the gaze rectification code data_processing_core.py
from this page before using our data processing code.
If you use the codes in this page, please cite our survey:
@article{Cheng2021Survey,
title={Appearance-based Gaze Estimation With Deep Learning: A Review and Benchmark},
author={Yihua Cheng and Haofei Wang and Yiwei Bao and Feng Lu},
journal={arXiv preprint arXiv:2104.12668},
year={2021}
}
GazeCapture
Data pre-processing
We follow the evaluation protocol of GazeCapture
.
They already split the whole dataset into training, test and validation set.
We crop the face and eye images based on the provided annotation.
Note that, the cropped image maybe have different size.
You need to resize them into suitable size in your code.
Codes of pre-processing
Please download here.
The code contains following parameters.
root = "/home/cyh/dataset/Original/GazeCapture"
out_root = "/home/cyh/dataset/FaceBased/GazeCapture"
- The
root
is the path of originalGazeCapture
dataset. - The
out_root
is the path for saving result file.
To use the code, you should first set the two paramters, and run
python data_processing_gazecapture.py
We also provide the guide for using our normalized data.
Dataset links
The original dataset can be downloaded from here.
To apply our code on this original dataset, please also cite:
@InProceedings{Krafka_2016_CVPR,
author = {Krafka, Kyle and Khosla, Aditya and Kellnhofer, Petr and Kannan, Harini and Bhandarkar, Suchendra and Matusik, Wojciech and Torralba, Antonio},
title = {Eye Tracking for Everyone},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2016}
}
MPIIFaceGaze
This dataset provides face and eye images. It also provides the PoG.
Data pre-processing
We follow the original protocol to split dataset. A total of 15 subjects and 3000 images of each subject is used for evaluation. We crop the face and eye images, and directly use the provided PoR in MPIIGaze.
Codes of pre-processing
Please download here. The code contains following parameters.
root = "/home/cyh/dataset/Original/MPIIFaceGaze"
sample_root = "/home/cyh/dataset/Original/MPIIGaze/Origin/Evaluation Subset/sample list for eye image"
out_root = "/home/cyh/dataset/GazePoint/MPIIGaze"
- The
root
is the path ofMPIIFaceGaze
. - The
sample_root
indicate the sample list inMPIIGaze
. Note that, this file is not contained inMPIIFaceGaze
. You should downloadMPIIGaze
for this file. - The
out_root
is the path for saving result file.
To use the code, you should set the three parameters first., and run
python data_processing_mpii.py
The processed data is saved in out_root
.
We also provide the guide for using our normalized data.
Dataset links
The original dataset can be downloaded from here.
To apply our code on this original dataset, please also cite:
@inproceedings{zhang2017s,
title={It’s written all over your face: Full-face appearance-based gaze estimation},
author={Zhang, Xucong and Sugano, Yusuke and Fritz, Mario and Bulling, Andreas},
booktitle={Computer Vision and Pattern Recognition Workshops (CVPRW), 2017 IEEE Conference on},
pages={2299--2308},
year={2017},
organization={IEEE}
}
EyeDiap
This dataset provides face and eye images. We also provide the PoR.
Data pre-processing
We select VGA videos of screen targets and sample one image per fifteen frames. We then truncate the sampled image of each video to ensure the same number of images. We crop the face and eye images from each frame, and use the provided PoR. In our benchmark, we randomly divide whole subjects into four clusters and perform four-fold validation.
Codes of pre-processing
Please download here. The code contains following parameters.
root = "/home/cyh/dataset/Original/EyeDiap/Data"
out_root = "/home/cyh/dataset/EyeBased/EyeDiap"
- The
root
is the path of originalEyeDiap
dataset. - The
out_root
is the path for saving result file.
To use the code, you should first set the two paramters, and run
python data_processing_diap.py
The normalized data is saved in out_root
.
We also provide the guide for using our normalized data.
please download here for cluster_diap.py
.
Dataset links
The original dataset can be downloaded from here.
To apply our code on this original dataset, please also cite:
@inproceedings{eyediap,
author = {Funes Mora, Kenneth Alberto and Monay, Florent and Odobez, Jean-Marc},
title = {EYEDIAP: A Database for the Development and Evaluation of Gaze Estimation Algorithms from RGB and RGB-D Cameras},
year = {2014},
isbn = {9781450327510},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
url = {https://doi.org/10.1145/2578153.2578190},
doi = {10.1145/2578153.2578190},
booktitle = {Proceedings of the Symposium on Eye Tracking Research and Applications},
pages = {255–258},
numpages = {4},
keywords = {natural-light, database, RGB-D, RGB, remote sensing, gaze estimation, depth, head pose},
location = {Safety Harbor, Florida},
series = {ETRA '14}
}