Enhancing efficiency and propulsion in bio-mimetic robotic fish through end-to-end deep reinforcement learning

被引:7
|
作者
Cui, Xinyu [1 ,2 ]
Sun, Boai [3 ]
Zhu, Yi [4 ,5 ]
Yang, Ning [1 ,2 ]
Zhang, Haifeng [1 ,2 ]
Cui, Weicheng [4 ,5 ]
Fan, Dixia [4 ,5 ]
Wang, Jun [6 ]
机构
[1] Inst Automat, Inst Automat, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing 100049, Peoples R China
[3] Zhejiang Univ, Zhejiang Univ Westlake Univ Joint Training, Hangzhou 310027, Peoples R China
[4] Westlake Univ, Sch Engn, Key Lab Coastal Environm & Resources Zhejiang Pro, Hangzhou 310030, Peoples R China
[5] Inst Adv Technol, Westlake Inst Adv Study, Hangzhou 310024, Peoples R China
[6] UCL, Dept Comp Sci, London WC1E 6BT, England
关键词
CARTESIAN-GRID SIMULATIONS; MULTIJOINT; OPTIMIZATION;
D O I
10.1063/5.0192993
中图分类号
O3 [力学];
学科分类号
08 ; 0801 ;
摘要
Aquatic organisms are known for their ability to generate efficient propulsion with low energy expenditure. While existing research has sought to leverage bio-inspired structures to reduce energy costs in underwater robotics, the crucial role of control policies in enhancing efficiency has often been overlooked. In this study, we optimize the motion of a bio-mimetic robotic fish using deep reinforcement learning (DRL) to maximize propulsion efficiency and minimize energy consumption. Our novel DRL approach incorporates extended pressure perception, a transformer model processing sequences of observations, and a policy transfer scheme. Notably, significantly improved training stability and speed within our approach allow for end-to-end training of the robotic fish. This enables agiler responses to hydrodynamic environments and possesses greater optimization potential compared to pre-defined motion pattern controls. Our experiments are conducted on a serially connected rigid robotic fish in a free stream with a Reynolds number of 6000 using computational fluid dynamics simulations. The DRL-trained policies yield impressive results, demonstrating both high efficiency and propulsion. The policies also showcase the agent's embodiment, skillfully utilizing its body structure and engaging with surrounding fluid dynamics, as revealed through flow analysis. This study provides valuable insights into the bio-mimetic underwater robots optimization through DRL training, capitalizing on their structural advantages, and ultimately contributing to more efficient underwater propulsion systems.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] End-to-End Deep Learning for Robotic Following
    Pierre, John M.
    ICMSCE 2018: PROCEEDINGS OF THE 2018 2ND INTERNATIONAL CONFERENCE ON MECHATRONICS SYSTEMS AND CONTROL ENGINEERING, 2015, : 77 - 85
  • [2] Robotic Odor Source Localization via End-to-End Recurrent Deep Reinforcement Learning
    Wang, Lingxiao
    Pang, Shuo
    2023 SEVENTH IEEE INTERNATIONAL CONFERENCE ON ROBOTIC COMPUTING, IRC 2023, 2023, : 43 - 50
  • [3] Towards end-to-end formation control for robotic fish via deep reinforcement learning with non-expert imitation?
    Sun, Yihao
    Yan, Chao
    Xiang, Xiaojia
    Zhou, Han
    Tang, Dengqing
    Zhu, Yi
    OCEAN ENGINEERING, 2023, 271
  • [4] End-to-End Robotic Reinforcement Learning without Reward Engineering
    Singh, Avi
    Yang, Larry
    Hartikainen, Kristian
    Finn, Chelsea
    Levine, Sergey
    ROBOTICS: SCIENCE AND SYSTEMS XV, 2019,
  • [5] NeuroVectorizer: End-to-End Vectorization with Deep Reinforcement Learning
    Haj-Ali, Ameer
    Ahmed, Nesreen K.
    Willke, Ted
    Shao, Yakun Sophia
    Asanovic, Krste
    Stoica, Ion
    CGO'20: PROCEEDINGS OF THE18TH ACM/IEEE INTERNATIONAL SYMPOSIUM ON CODE GENERATION AND OPTIMIZATION, 2020, : 242 - 255
  • [6] End-to-End Deep Reinforcement Learning for Exoskeleton Control
    Rose, Lowell
    Bazzocchi, Michael C. F.
    Nejat, Goldie
    2020 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2020, : 4294 - 4301
  • [7] End-to-End Race Driving with Deep Reinforcement Learning
    Jaritz, Maximilian
    de Charette, Raoul
    Toromanoff, Marin
    Perot, Etienne
    Nashashibi, Fawzi
    2018 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2018, : 2070 - 2075
  • [8] End-to-End Deep Reinforcement Learning for Conversation Disentanglement
    Bhukar, Karan
    Kumar, Harshit
    Raghu, Dinesh
    Gupta, Ajay
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 11, 2023, : 12571 - 12579
  • [9] End-to-end Reinforcement Learning of Robotic Manipulation with Robust Keypoints Representation
    Wang, Tianying
    Puang, En Yen
    Lee, Marcus
    Jing, Wei
    Wu, Yan
    PROCEEDINGS OF 2022 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2022, : 590 - 597
  • [10] End-to-end Control of Kart Agent with Deep Reinforcement Learning
    Zhang Ruiming
    Liu Chengju
    Chen Qijun
    2018 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (ROBIO), 2018, : 1688 - 1693