谷歌浏览器插件
订阅小程序
在清言上使用

Bridging Visual Perception With Contextual Semantics For Understanding Robot Manipulation Tasks

2020 IEEE 16TH INTERNATIONAL CONFERENCE ON AUTOMATION SCIENCE AND ENGINEERING (CASE)(2020)

引用 2|浏览28
暂无评分
摘要
Understanding manipulation scenarios allows intelligent robots to plan for appropriate actions to complete a manipulation task successfully. It is essential for intelligent robots to semantically interpret manipulation knowledge by describing entities, relations and attributes in a structural manner. In this paper, we propose an implementing framework to generate high-level conceptual dynamic knowledge graphs from video clips. A combination of a Vision-Language model and an ontology system, in correspondence with visual perception and contextual semantics, is used to represent robot manipulation knowledge with Entity-Relation-Entity (E-R-E) and Entity-Attribute-Value (E-A-V) tuples. The proposed method is flexible and well-versed. Using the framework, we present a case study where robot performs manipulation actions in a kitchen environment, bridging visual perception with contextual semantics using the generated dynamic knowledge graphs.
更多
查看译文
关键词
robot manipulation tasks,generated dynamic knowledge graphs,visual perception,manipulation actions,Entity-Relation-Entity,robot manipulation knowledge,contextual semantics,Vision-Language model,high-level conceptual dynamic knowledge graphs,manipulation task,intelligent robots
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要