Perceptual User Interfaces Logo
University of Stuttgart Logo

Visual Intention Classification by Deep Learning for Gaze-based Human-Robot Interaction

Lei Shi, Cosmin Copot, Steve Vanlanduit

IFAC-PapersOnLine, 53(5), pp. 750-755, 2020.


Abstract

In this work, we propose a deep learning model to classify a human’s visual intention in gaze-based Human-Robot Interaction(HRI). We consider a scenario in which a human wears a pair of eye tracking glasses and can select an object by gaze and a robotic manipulator picks up the object. A neural network is trained as a binary classifier to classify if a human is looking at an object. The network architecture is based on Fully Convolutional Net(FCN), Convolutional Block Attention Modules(CBAM) and Residual Blocks. We evaluate our model with two experiments. In one experiment we test the performance in the scenario where only a single object exists and the other one multiple objects exist. The results show that our proposed network is accurate and it can generalize well. The F1 score on the single object is 0.971 and 0.962 on multiple objects.

Links


BibTeX

@article{shi20_ifac, author = {Shi, Lei and Copot, Cosmin and Vanlanduit, Steve}, title = {Visual Intention Classification by Deep Learning for Gaze-based Human-Robot Interaction}, journal = {IFAC-PapersOnLine}, year = {2020}, volume = {53}, number = {5}, pages = {750-755}, doi = {10.1016/j.ifacol.2021.04.168} }