GAze on TArget (GATA) dataset is a large-scale annotated gaze dataset, tailored for training deep learning architectures. It was created following the “target search” paradigm where subjects were asked to visually search for a specific object class. Forty eight different subjects participated in the recording procedure using myGaze capturing sensor.
Figure 1: Gaze annotation process
Figure 2: Various classes heatmaps
Figure 3: Visualization of the implicit human response: fixation scan-path (left) and the corresponding heatmap (right)
The gaze annotations are provided with the json file format using the following naming convention. objectID_imageID.json The first part (objectID) denotes the target object class id and the second part (imageID) the COCO image id respectively. Inside the json file gaze points annotations, time stamp and x,y coordinates, are included as presented below. [{"Time":12510830749,"X":343,"Y":285}, {"Time":12510864083,"X":343,"Y":285}, {"Time":12510897394,"X":343,"Y":285}, {"Time":12510930745,"X":343,"Y":286}, {"Time":12510964081,"X":343,"Y":287}]
The proposed dataset was utilized for building a deep learning model capable of predicting objects in an image as relevant or non-relevant, based on gaze, according to the users’ preferences. The proposed model is presented below:
Figure 4: The proposed DL architecture for user relevance assessment, where DEi is the distance embedding for ith fixation and Oj is the jth object class.
You can download the dataset here.
Stavridis, K., Psaltis, A., Dimou, A., Papadopoulos G. Th., & Daras, P. (2019). "Deep Spatio-Temporal Modeling for Object-Level Gaze-Based Relevance Assessment", In 2019 27th European Signal Processing Conference (EUSIPCO). IEEE.