Split Deep Q-Learning for Robust Object Singulation

被引:0
|
作者
Sarantopoulos, Iason [1 ]
Kiatos, Marios [1 ,2 ]
Doulgeri, Zoe [1 ]
Malassiotis, Sotiris [2 ]
机构
[1] Aristotle Univ Thessaloniki, Dept Elect & Comp Engn, Thessaloniki 54124, Greece
[2] Informat Technol Inst ITI, Ctr Res & Technol Hellas CERTH, Thessaloniki 57001, Greece
基金
欧盟地平线“2020”;
关键词
D O I
10.1109/icra40945.2020.9196647
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Extracting a known target object from a pile of other objects in a cluttered environment is a challenging robotic manipulation task encountered in many robotic applications. In such conditions, the target object touches or is covered by adjacent obstacle objects, thus rendering traditional grasping techniques ineffective. In this paper, we propose a pushing policy aiming at singulating the target object from its surrounding clutter, by means of lateral pushing movements of both the neighboring objects and the target object until sufficient 'grasping room' has been achieved. To achieve the above goal we employ reinforcement learning and particularly Deep Q-learning (DQN) to learn optimal push policies by trial and error. A novel Split DQN is proposed to improve the learning rate and increase the modularity of the algorithm. Experiments show that although learning is performed in a simulated environment the transfer of learned policies to a real environment is effective thanks to robust feature selection. Finally, we demonstrate that the modularity of the algorithm allows the addition of extra primitives without retraining the model from scratch.
引用
收藏
页码:6225 / 6231
页数:7
相关论文
共 50 条
  • [21] On-Off Adversarially Robust Q-Learning
    Sahoo, Prachi Pratyusha
    Vamvoudakis, Kyriakos G.
    IEEE CONTROL SYSTEMS LETTERS, 2020, 4 (03): : 749 - 754
  • [22] Experience-Based Heuristic Search: Robust Motion Planning with Deep Q-Learning
    Bernhard, Julian
    Gieselmann, Robert
    Esterle, Klemens
    Knoll, Alois
    2018 21ST INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2018, : 3175 - 3182
  • [23] Deep Q-Learning with Phased Experience Cooperation
    Wang, Hongbo
    Zeng, Fanbing
    Tu, Xuyan
    COMPUTER SUPPORTED COOPERATIVE WORK AND SOCIAL COMPUTING, CHINESECSCW 2019, 2019, 1042 : 752 - 765
  • [24] Stochastic Variance Reduction for Deep Q-learning
    Zhao, Wei-Ye
    Peng, Jian
    AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 2318 - 2320
  • [25] Deep Surrogate Q-Learning for Autonomous Driving
    Kalweit, Maria
    Kalweit, Gabriel
    Werling, Moritz
    Boedecker, Joschka
    2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2022), 2022, : 1578 - 1584
  • [26] Trading ETFs with Deep Q-Learning Algorithm
    Hong, Shao-Yan
    Liu, Chien-Hung
    Chen, Woei-Kae
    You, Shingchern D.
    2020 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS - TAIWAN (ICCE-TAIWAN), 2020,
  • [27] Diagnosing Bottlenecks in Deep Q-learning Algorithms
    Fu, Justin
    Kumar, Aviral
    Soh, Matthew
    Levine, Sergey
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [28] Deep Q-Learning for Aggregator Price Design
    Pigott, Aisling
    Baker, Kyri
    Mosiman, Cory
    2021 IEEE POWER & ENERGY SOCIETY GENERAL MEETING (PESGM), 2021,
  • [29] NeuroHex: A Deep Q-learning Hex Agent
    Young, Kenny
    Vasan, Gautham
    Hayward, Ryan
    COMPUTER GAMES: 5TH WORKSHOP ON COMPUTER GAMES, CGW 2016, AND 5TH WORKSHOP ON GENERAL INTELLIGENCE IN GAME-PLAYING AGENTS, GIGA 2016, HELD IN CONJUNCTION WITH THE 25TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2016, NEW YORK, USA, JULY 9-10, 2016, 2017, 705 : 3 - 18
  • [30] QLP: Deep Q-Learning for Pruning Deep Neural Networks
    Camci, Efe
    Gupta, Manas
    Wu, Min
    Lin, Jie
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (10) : 6488 - 6501