收录:
摘要:
It is hoped that the robot could interact with the human when the robots help us in our daily lives. And understanding humans’ specific intention is the first crucial task for human-robot interaction. In this paper, we firstly develop a multi-task model for recognizing humans’ intention, which is composed of two sub-tasks: human action recognition and hand-held object identification. For the front subtask, an effective ST-GCN-LSTM model is proposed by fusing the Spatial Temporal Graph Convolutional Networks and Long Short Term Memory Networks. And for the second subtask, the YOLO v3 model is adopted for the hand-held object identification. Then, we build a framework for robot interacting with the human. Finally, these proposed models and the interacting framework are verified on several datasets and the testing results show the effectiveness of the proposed models and the framework. © 2020 Elsevier B.V.
关键词:
通讯作者信息:
电子邮件地址:
来源 :
Neurocomputing
ISSN: 0925-2312
年份: 2021
卷: 430
页码: 174-184
6 . 0 0 0
JCR@2022
ESI学科: COMPUTER SCIENCE;
ESI高被引阀值:87
JCR分区:2
归属院系: