Abstract:Existing public eye-tracking datasets almost entirely lack data collection for the deaf population, exhibit unbalanced eye movement type distributions, and rarely include large-scale datasets with comprehensive labeling. To address these issues, this paper proposes and constructs the Eye Tracking for Deaf Aerobics (ETFDA) dataset. First, using eye-tracking technology, data from 112 deaf and 54 hearing individuals are collected to fill the gap in deaf data. Second, aerobics videos are recorded as stimulus materials to induce saccadic and smooth pursuit eye movements, addressing the unbalanced distribution of eye movement types. Finally, through preprocessing, feature extraction, and annotation, a dataset containing approximately 1.33 million records covering four types of eye movement labels is constructed to enrich existing eye-tracking dataset resources. Based on this dataset, comparing the eye movement characteristics of deaf and hearing individuals reveals significant differences, providing new insights into understanding the visual cognitive mechanisms of the deaf. Additionally, through comparative analysis and algorithm validation with other public datasets, the results show that this dataset has a more balanced distribution of eye movement types and more refined and effective labels, offering significant application value.