Fish growth trajectory tracking using Q-learning in precision aquaculture

被引:15
|
作者
Chahid, Abderrazak [1 ]
N'Doye, Ibrahima [1 ]
Majoris, John E. [2 ]
Berumen, Michael L. [2 ]
Laleg-Kirati, Taous-Meriem [1 ]
机构
[1] King Abdullah Univ Sci & Technol KAUST, Elect & Math Sci & Engn Div CEMSE, Thuwal 239556900, Makkah Province, Saudi Arabia
[2] King Abdullah Univ Sci & Technol KAUST, Red Sea Res Ctr, Biol & Environm Sci & Engn Div, Thuwal 239556900, Makkah Province, Saudi Arabia
关键词
Fish growth model; Reference trajectory tracking; Markov decision process; Process control; Q-learning; Reinforcement learning; TILAPIA OREOCHROMIS-NILOTICUS; MODEL; OUTPUT; POND;
D O I
10.1016/j.aquaculture.2021.737838
中图分类号
S9 [水产、渔业];
学科分类号
0908 ;
摘要
This paper studies fish growth trajectory tracking using Q-learning under a representative bioenergetic growth model of Nile tilapia (Oreochromis niloticus). The fish growth rate varies in practice and cannot be easily estimated due to the complex aquaculture condition and variable environmental factors. Additionally, the growth trajectory tracking problem is challenging to solve by most of the model-based control approaches due to the nonlinear couplings and interactions between multi-inputs such as temperature, dissolved oxygen, un-ionized ammonia, and the model uncertainty of the fish growth system. We formulate the growth trajectory tracking problem as sampled-data optimal control using discrete state-action pairs Markov decision process on the simulated growth trajectories data to mimic the real aquaculture environment adequately. We propose two Q-learning algorithms that learn the optimal control policy from the simulated data of the fish growth trajectories beginning from the juvenile stage until the desired market weight in the aquaculture environment. The first Q-learning scheme learns the optimal feeding control policy to fish growth rate cultured in cages, while the second one online updates the optimal feeding control policy within an optimal temperature profile for the aquaculture fish growth rate in tanks. The simulation results demonstrate that both Q-learning control strategies achieve good trajectory tracking performance with lower feeding rates and help compensate for the environmental changes of the manipulated variables and the bioenergetic model uncertainties of fish growth in the aquaculture environment. The proposed Q-learning control policies achieve 1.7% and 6.6% relative trajectory tracking errors of the average total weight of fish from both tanks on land and floating cages, respectively. Furthermore, the feeding and temperature control policies reduce 11% relative feeding quantity of the food waste in tanks on land compared to the floating cages where the water temperature is maintained at the ambient temperature of 29.7 degrees C.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] Optimal Trajectory Output Tracking Control with a Q-learning Algorithm
    Vamvoudakis, Kyriakos G.
    2016 AMERICAN CONTROL CONFERENCE (ACC), 2016, : 5752 - 5757
  • [2] Model predictive control paradigms for fish growth reference tracking in precision aquaculture
    Chahid, Abderrazak
    N'Doye, Ibrahima
    Majoris, John E.
    Berumen, Michael L.
    Laleg-Kirati, Taous Meriem
    JOURNAL OF PROCESS CONTROL, 2021, 105 : 160 - 168
  • [3] Model predictive control paradigms for fish growth reference tracking in precision aquaculture
    Chahid, Abderrazak
    N'Doye, Ibrahima
    Majoris, John E.
    Berumen, Michael L.
    Laleg-Kirati, Taous Meriem
    Journal of Process Control, 2021, 105 : 160 - 168
  • [4] Tuna Nutriment Tracking using Trajectory Mapping in Application to Aquaculture Fish Tank
    Pradana, Hilmil
    Horio, Keiichi
    2020 DIGITAL IMAGE COMPUTING: TECHNIQUES AND APPLICATIONS (DICTA), 2020,
  • [5] Multi Target Tracking using a Compact Q-Learning with a Teacher
    Saad, E. M.
    Awadalla, M. H.
    Hamdy, A. M.
    Ali, H. I.
    ICCES: 2008 INTERNATIONAL CONFERENCE ON COMPUTER ENGINEERING & SYSTEMS, 2007, : 173 - 178
  • [6] Adaptive-Precision Framework for SGD Using Deep Q-Learning
    Zhang, Wentai
    Huang, Hanxian
    Zhang, Jiaxi
    Jiang, Ming
    Luo, Guojie
    2018 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER-AIDED DESIGN (ICCAD) DIGEST OF TECHNICAL PAPERS, 2018,
  • [7] Optimal trajectory tracking for uncertain linear discrete-time systems using time-varying Q-learning
    Geiger, Maxwell
    Narayanan, Vignesh
    Jagannathan, Sarangapani
    INTERNATIONAL JOURNAL OF ADAPTIVE CONTROL AND SIGNAL PROCESSING, 2024, 38 (07) : 2340 - 2368
  • [8] Output Feedback Optimal Tracking Control Using Reinforcement Q-Learning
    Rizvi, Syed Ali Asad
    Lin, Zongli
    2018 ANNUAL AMERICAN CONTROL CONFERENCE (ACC), 2018, : 3423 - 3428
  • [9] Multi-Target Tracking using a Compact Q-Learning with a Teacher
    Saad, E. M.
    Awadalla, M. H.
    Hamdy, A. M.
    Ali, H. I.
    NRSC: 2009 NATIONAL RADIO SCIENCE CONFERENCE: NRSC 2009, VOLS 1 AND 2, 2009, : 284 - 295
  • [10] Maneuvering Target Tracking Using Q-learning Based Kalman Filter
    Bekhtaoui, Z.
    Meche, A.
    Dahmani, M.
    Meraim, K. Abed
    2017 5TH INTERNATIONAL CONFERENCE ON ELECTRICAL ENGINEERING - BOUMERDES (ICEE-B), 2017,