A. Ramos,Javier Muguerza,David López Pérez,Unai Elordi,Luis Unzueta,Arantxa Villanueva
标识
DOI:10.1145/3649902.3656367
摘要
This paper aims to advance the fields of unsupervised data labeling and incremental cross-domain training techniques. We apply these innovative methods to develop a model tailored for the Augmentative and Alternative Communication (AAC) application domain, introducing a new perspective in hybrid eye Gaze Estimation (GE). These hybrid eye GE models combine the generalization strengths of appearance-based models with the scene understanding capabilities inherent in geometrical reconstruction. The use of open eye tracking datasets for the AAC domain introduces domain shift, while accurately labeling gaze vectors is challenging without specialized hardware for proper 3D dimensional reconstruction. We propose an approach to solve this challenges by conducting standardized unsupervised gaze vector labeling across multiple open GE datasets and subsequently performing incremental training to adapt to the target domain. Using a proprietary dataset we were able to reduce the gaze error from 4.87º to 3.95º, compared to a traditional single-step training.