Fish growth trajectory tracking using Q-learning in precision aquaculture

被引:15
|
作者
Chahid, Abderrazak [1 ]
N'Doye, Ibrahima [1 ]
Majoris, John E. [2 ]
Berumen, Michael L. [2 ]
Laleg-Kirati, Taous-Meriem [1 ]
机构
[1] King Abdullah Univ Sci & Technol KAUST, Elect & Math Sci & Engn Div CEMSE, Thuwal 239556900, Makkah Province, Saudi Arabia
[2] King Abdullah Univ Sci & Technol KAUST, Red Sea Res Ctr, Biol & Environm Sci & Engn Div, Thuwal 239556900, Makkah Province, Saudi Arabia
关键词
Fish growth model; Reference trajectory tracking; Markov decision process; Process control; Q-learning; Reinforcement learning; TILAPIA OREOCHROMIS-NILOTICUS; MODEL; OUTPUT; POND;
D O I
10.1016/j.aquaculture.2021.737838
中图分类号
S9 [水产、渔业];
学科分类号
0908 ;
摘要
This paper studies fish growth trajectory tracking using Q-learning under a representative bioenergetic growth model of Nile tilapia (Oreochromis niloticus). The fish growth rate varies in practice and cannot be easily estimated due to the complex aquaculture condition and variable environmental factors. Additionally, the growth trajectory tracking problem is challenging to solve by most of the model-based control approaches due to the nonlinear couplings and interactions between multi-inputs such as temperature, dissolved oxygen, un-ionized ammonia, and the model uncertainty of the fish growth system. We formulate the growth trajectory tracking problem as sampled-data optimal control using discrete state-action pairs Markov decision process on the simulated growth trajectories data to mimic the real aquaculture environment adequately. We propose two Q-learning algorithms that learn the optimal control policy from the simulated data of the fish growth trajectories beginning from the juvenile stage until the desired market weight in the aquaculture environment. The first Q-learning scheme learns the optimal feeding control policy to fish growth rate cultured in cages, while the second one online updates the optimal feeding control policy within an optimal temperature profile for the aquaculture fish growth rate in tanks. The simulation results demonstrate that both Q-learning control strategies achieve good trajectory tracking performance with lower feeding rates and help compensate for the environmental changes of the manipulated variables and the bioenergetic model uncertainties of fish growth in the aquaculture environment. The proposed Q-learning control policies achieve 1.7% and 6.6% relative trajectory tracking errors of the average total weight of fish from both tanks on land and floating cages, respectively. Furthermore, the feeding and temperature control policies reduce 11% relative feeding quantity of the food waste in tanks on land compared to the floating cages where the water temperature is maintained at the ambient temperature of 29.7 degrees C.
引用
收藏
页数:9
相关论文
共 50 条
  • [21] Optimal Tracking Control of Nonlinear Multiagent Systems Using Internal Reinforce Q-Learning
    Peng, Zhinan
    Luo, Rui
    Hu, Jiangping
    Shi, Kaibo
    Nguang, Sing Kiong
    Ghosh, Bijoy Kumar
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (08) : 4043 - 4055
  • [22] Multi-UAV mmWave Beam Tracking Using Q-Learning and Interference Mitigation
    Chiang, Hsiao-Lan
    Chen, Kwang-Chong
    Rave, Wolfgang
    Marandi, Mostafa Khalili
    Fettweis, Gerhard
    2020 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS WORKSHOPS (ICC WORKSHOPS), 2020,
  • [23] LONG AND SHORT MEMORY BALANCING IN VISUAL CO-TRACKING USING Q-LEARNING
    Meshgi, Kourosh
    Mirzaei, Maryam Sadat
    Oba, Shigeyuki
    2019 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2019, : 3970 - 3974
  • [24] Random Graphs Estimation using Q-Learning
    Babahaji, Mina
    Blouin, Stephane
    Lucia, Walter
    Asadi, M. Mehdi
    Mahboubi, Hamid
    Aghdam, Amir G.
    2021 IEEE INTERNATIONAL CONFERENCE ON WIRELESS FOR SPACE AND EXTREME ENVIRONMENTS (WISEE), 2021,
  • [25] Intelligent transportation system using Q-learning
    Park, MS
    Kim, PJ
    Choi, JY
    2003 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS, VOLS 1-5, CONFERENCE PROCEEDINGS, 2003, : 4684 - 4687
  • [26] Automated Portfolio Rebalancing using Q-learning
    Darapaneni, Narayana
    Basu, Amitavo
    Savla, Sanket
    Gururajan, Raamanathan
    Saquib, Najmus
    Singhavi, Sudarshan
    Kale, Aishwarya
    Bid, Pratik
    Paduri, Anwesh Reddy
    2020 11TH IEEE ANNUAL UBIQUITOUS COMPUTING, ELECTRONICS & MOBILE COMMUNICATION CONFERENCE (UEMCON), 2020, : 596 - 602
  • [27] Robot behavioral selection using Q-learning
    Martinson, E
    Stoytchev, A
    Arkin, R
    2002 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, VOLS 1-3, PROCEEDINGS, 2002, : 970 - 977
  • [28] Algorithmic Forex Trading Using Q-learning
    Zahrah, Hasna Haifa
    Tirtawangsa, Jimmy
    ARTIFICIAL INTELLIGENCE APPLICATIONS AND INNOVATIONS, AIAI 2023, PT I, 2023, 675 : 24 - 35
  • [29] Adaptive visual tracking using the prioritized Q-learning algorithm: MDP-based parameter learning approach
    Khim, Sarang
    Hong, Sungjin
    Kim, Yoonyoung
    Rhee, Phill Kyu
    IMAGE AND VISION COMPUTING, 2014, 32 (12) : 1090 - 1101
  • [30] Multiagent Q-learning based UAV trajectory planning for effective situational awareness
    Akin, Erdal
    Demir, Kubilay
    Yetgin, Halil
    TURKISH JOURNAL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCES, 2021, 29 (05) : 2561 - 2579