Optimising maize threshing process with temporal proximity soft actor-critic deep reinforcement learning algorithm

被引:0
|
作者
Zhang, Qiang [1 ]
Fang, Xuwen [1 ]
Gao, Xiaodi [1 ,2 ]
Zhang, Jinsong [1 ]
Zhao, Xuelin [1 ]
Yu, Lulu [1 ]
Yu, Chunsheng [1 ]
Zhou, Deyi [1 ]
Zhou, Haigen [1 ]
Zhang, Li [1 ]
Wu, Xinling [1 ]
机构
[1] Jilin Univ, Coll Biol & Agr Engn, Changchun 130022, Peoples R China
[2] Jilin Jianzhu Univ, Sch Emergency Sci & Engn, Changchun 130118, Peoples R China
关键词
Threshing quality optimisation; Agricultural machinery; Machine learning; Agricultural automation; Sensitivity analysis; DAMAGE;
D O I
10.1016/j.biosystemseng.2024.11.001
中图分类号
S2 [农业工程];
学科分类号
0828 ;
摘要
Maize threshing is a crucial process in grain production, and optimising it is essential for reducing post-harvest losses. This study proposes a model-based temporal proximity soft actor-critic (TP-SAC) algorithm to optimise the maize threshing process in the threshing drum. The proposed approach employs an LSTM model as a real-time predictor of threshing quality, achieving an R2 of 97.17% and 98.43% for damage and unthreshed rates on the validation set. In actual threshing experiments, the LSTM model demonstrates an average error of 5.45% and 3.83% for damage and unthreshed rates. The LSTM model is integrated with the TP-SAC algorithm, acting as the environment with which the TP-SAC interacts, enabling efficient training with limited real-world data. The TPSAC algorithm addresses the temporal correlation in the threshing process by incorporating temporal proximity sampling into the SAC algorithm's experience replay mechanism. TP-SAC outperforms the standard SAC algorithm in the simulated environment, demonstrating better sample efficiency and faster convergence. When deployed in actual threshing operations, the TP-SAC algorithm reduces the damage rate by an average of 0.91% across different feed rates compared to constant control. The proposed TP-SAC algorithm offers a novel and practical approach to optimising the maize threshing process, enhancing threshing quality.
引用
收藏
页码:229 / 239
页数:11
相关论文
共 50 条
  • [21] A Deep Actor-Critic Reinforcement Learning Framework for Dynamic Multichannel Access
    Zhong, Chen
    Lu, Ziyang
    Gursoy, M. Cenk
    Velipasalar, Senem
    IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2019, 5 (04) : 1125 - 1139
  • [22] A World Model for Actor-Critic in Reinforcement Learning
    Panov, A. I.
    Ugadiarov, L. A.
    PATTERN RECOGNITION AND IMAGE ANALYSIS, 2023, 33 (03) : 467 - 477
  • [23] Curious Hierarchical Actor-Critic Reinforcement Learning
    Roeder, Frank
    Eppe, Manfred
    Nguyen, Phuong D. H.
    Wermter, Stefan
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2020, PT II, 2020, 12397 : 408 - 419
  • [24] Actor-Critic based Improper Reinforcement Learning
    Zaki, Mohammadi
    Mohan, Avinash
    Gopalan, Aditya
    Mannor, Shie
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [25] Applying Online Expert Supervision in Deep Actor-Critic Reinforcement Learning
    Zhang, Jin
    Chen, Jiansheng
    Huang, Yiqing
    Wan, Weitao
    Li, Tianpeng
    PATTERN RECOGNITION AND COMPUTER VISION, PT II, 2018, 11257 : 469 - 478
  • [26] AN ACTOR-CRITIC REINFORCEMENT LEARNING ALGORITHM BASED ON ADAPTIVE RBF NETWORK
    Li, Chun-Gui
    Wang, Meng
    Huang, Zhen-Jin
    Zhang, Zeng-Fang
    PROCEEDINGS OF 2009 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-6, 2009, : 984 - 988
  • [27] Coverage Path Planning Using Actor-Critic Deep Reinforcement Learning
    Garrido-Castaneda, Sergio Isahi
    Vasquez, Juan Irving
    Antonio-Cruz, Mayra
    SENSORS, 2025, 25 (05)
  • [28] Fully distributed actor-critic architecture for multitask deep reinforcement learning
    Valcarcel Macua, Sergio
    Davies, Ian
    Tukiainen, Aleksi
    De Cote, Enrique Munoz
    KNOWLEDGE ENGINEERING REVIEW, 2021, 36
  • [29] A fuzzy Actor-Critic reinforcement learning network
    Wang, Xue-Song
    Cheng, Yu-Hu
    Yi, Jian-Qiang
    INFORMATION SCIENCES, 2007, 177 (18) : 3764 - 3781
  • [30] Research on actor-critic reinforcement learning in RoboCup
    Guo, He
    Liu, Tianying
    Wang, Yuxin
    Chen, Feng
    Fan, Jianming
    WCICA 2006: SIXTH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-12, CONFERENCE PROCEEDINGS, 2006, : 205 - 205