Detecting Target Objects by Natural Language Instructions Using an RGB-D Camera

被引:2
|
作者
Bao, Jiatong [1 ]
Jia, Yunyi [2 ]
Cheng, Yu [3 ]
Tang, Hongru [1 ]
Xi, Ning [3 ]
机构
[1] Yangzhou Univ, Dept Hydraul Energy & Power Engn, Yangzhou 225127, Jiangsu, Peoples R China
[2] Clemson Univ, Dept Automot Engn, Greenville, SC 29607 USA
[3] Michigan State Univ, Dept Elect & Comp Engn, E Lansing, MI 48824 USA
来源
SENSORS | 2016年 / 16卷 / 12期
关键词
object grounding; target object detection; object recognition; natural language processing; natural language control; robotic manipulation system; FRAMEWORK;
D O I
10.3390/s16122117
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Controlling robots by natural language (NL) is increasingly attracting attention for its versatility, convenience and no need of extensive training for users. Grounding is a crucial challenge of this problem to enable robots to understand NL instructions from humans. This paper mainly explores the object grounding problem and concretely studies how to detect target objects by the NL instructions using an RGB-D camera in robotic manipulation applications. In particular, a simple yet robust vision algorithm is applied to segment objects of interest. With the metric information of all segmented objects, the object attributes and relations between objects are further extracted. The NL instructions that incorporate multiple cues for object specifications are parsed into domain-specific annotations. The annotations from NL and extracted information from the RGB-D camera are matched in a computational state estimation framework to search all possible object grounding states. The final grounding is accomplished by selecting the states which have the maximum probabilities. An RGB-D scene dataset associated with different groups of NL instructions based on different cognition levels of the robot are collected. Quantitative evaluations on the dataset illustrate the advantages of the proposed method. The experiments of NL controlled object manipulation and NL-based task programming using a mobile manipulator show its effectiveness and practicability in robotic applications.
引用
收藏
页数:23
相关论文
共 50 条
  • [31] Obstacle Avoidance for Mobile Robot using RGB-D Camera
    Thapa, Vikas
    Capoor, Shiv
    Sharma, Paawan
    Mondal, Amit Kumar
    PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON INTELLIGENT SUSTAINABLE SYSTEMS (ICISS 2017), 2017, : 1082 - 1087
  • [32] Using a RGB-D camera for 6DoF SLAM
    Munoz, Jose
    Pastor, Daniel
    Gil, Pablo
    Puente, Santiago
    Cazorla, Miguel
    ARTIFICIAL INTELLIGENCE RESEARCH AND DEVELOPMENT, 2012, 248 : 143 - +
  • [33] Human Pose Recognition and tracking using RGB-D Camera
    Moussaoui, Imene
    Boubertakh, Hamid
    PROCEEDINGS OF 2016 8TH INTERNATIONAL CONFERENCE ON MODELLING, IDENTIFICATION & CONTROL (ICMIC 2016), 2016, : 515 - 519
  • [34] Ambient Assistive Living System Using RGB-D Camera
    Daher, Mohamad
    El Najjar, Maan El Badaoui
    Diab, Ahmad
    Khalil, Mohamad
    Dib, Abdallah
    Charpillet, Francois
    2017 FOURTH INTERNATIONAL CONFERENCE ON ADVANCES IN BIOMEDICAL ENGINEERING (ICABME), 2017, : 17 - 20
  • [35] An Accurate Algorithm for the Identification of Fingertips Using an RGB-D Camera
    Maisto, Marco
    Panella, Massimo
    Liparulo, Luca
    Proietti, Andrea
    IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2013, 3 (02) : 272 - 283
  • [36] Visual Odometry using RGB-D Camera on Ceiling Vision
    Wang, Han
    Mou, Wei
    Suratno, Hendra
    Seet, Gerald
    Li, Maohai
    Lau, M. W. S.
    Wang, Danwei
    2012 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (ROBIO 2012), 2012,
  • [37] Kinematic Calibration of Robot Manipulator using RGB-D Camera
    Jang W.B.
    Lee J.
    Park S.H.
    Chung S.Y.
    Jin M.
    Hwang M.J.
    Journal of Institute of Control, Robotics and Systems, 2023, 29 (03) : 264 - 271
  • [38] A Novel Hybrid Visual Odometry Using an RGB-D Camera
    Wang, Huiguo
    Wu, Xinyu
    Chen, Zhiheng
    He, Yong
    PROCEEDINGS 2018 33RD YOUTH ACADEMIC ANNUAL CONFERENCE OF CHINESE ASSOCIATION OF AUTOMATION (YAC), 2018, : 47 - 51
  • [39] Sparse Edge Visual Odometry using an RGB-D Camera
    Hsu, Jhih-Lei
    Lin, Huei-Yung
    2017 11TH ASIAN CONTROL CONFERENCE (ASCC), 2017, : 964 - 969
  • [40] Semantic Mapping for Smart Wheelchairs Using RGB-D Camera
    Wei, Zhixuan
    Chen, Weidong
    Wang, Jingchuan
    JOURNAL OF MEDICAL IMAGING AND HEALTH INFORMATICS, 2013, 3 (01) : 94 - 100