A Novel Architectural Method for Producing Dynamic Gaze Behavior in Human-Robot Interactions

2022

Conference: Proceedings of the 2022 ACM/IEEE International Conference on Human-Robot Interaction
Pages: 383--392

Briggs, Gordon and Chita-Tegmark, Meia and Krause, Evan and Bridewell, Will and Bello, Paul and Scheutz, Matthias

We present a novel integration between a computational framework for modeling attention-driven perception and cognition (ARCADIA) with a cognitive robotic architecture (DIARC), demonstrating how this integration can be used to drive the gaze behavior of a robotic platform. Although some previous approaches to controlling gaze behavior in robots during human-robot interactions have relied either on models of human visual attention or human cognition, ARCADIA provides a novel framework with an attentional mechanism that bridges both lower-level visual and higher-level cognitive processes. We demonstrate how this approach can produce more natural and human-like robot gaze behavior. In particular, we focus on how our approach can control gaze during an interactive object learning task. We present results from a pilot crowdsourced evaluation that investigates whether the gaze behavior produced during this task increases confidence that the robot has correctly learned each object.

@inproceedings{briggsetal22hri,
  title={A Novel Architectural Method for Producing Dynamic Gaze Behavior in Human-Robot Interactions},
  author={Briggs, Gordon and Chita-Tegmark, Meia and Krause, Evan and
                  Bridewell, Will and Bello, Paul and Scheutz,
                  Matthias},
  year={2022},
  booktitle={Proceedings of the 2022 ACM/IEEE International Conference
                  on Human-Robot Interaction},
  pages={383--392}
  url={https://hrilab.tufts.edu/publications/briggsetal22hri.pdf}
  doi={10.1109/HRI53351.2022.9889499}
}