Robust Unsupervised Gaze Calibration Using Conversation and Manipulation Attention Priors

ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS(2022)

引用 1|浏览10
暂无评分
摘要
Gaze estimation is a difficult task, even for humans. However, as humans, we are good at understanding a situation and exploiting it to guess the expected visual focus of attention of people, and we usually use this information to retrieve people's gaze. In this article, we propose to leverage such situation-based expectation about people's visual focus of attention to collect weakly labeled gaze samples and perform person-specific calibration of gaze estimators in an unsupervised and online way. In this context, our contributions are the following: (i) we show how task contextual attention priors can be used to gather reference gaze samples, which is a cumbersome process otherwise; (ii) we propose a robust estimation framework to exploit these weak labels for the estimation of the calibration model parameters; and (iii) we demonstrate the applicability of this approach on two human-human and human-robot interaction settings, namely conversation and manipulation. Experiments on three datasets validate our approach. providing insights on the priors effectiveness and on the impact of different calibration models, particularly the usefulness of taking head pose into account.
更多
查看译文
关键词
Gaze estimation,visual focus of attention,remote sensor,RGB-D camera,conversation,manipulation,unsupervised calibration,online calibration
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要