Channel randomisation: Self-supervised representation learning for reliable visual anomaly detection in speciality crops

被引:1
|
作者
Choi, Taeyeong [1 ]
Would, Owen [2 ]
Salazar-Gomez, Adrian [2 ]
Liu, Xin [3 ]
Cielniak, Grzegorz [2 ]
机构
[1] Kennesaw State Univ, Dept Informat Technol, 1100 South Marietta Pkwy, Marietta, GA 30060 USA
[2] Univ Lincoln, Lincoln Inst Agrifood Technol, Riseholme Pk LN2 2LG, Lincoln, England
[3] Univ Calif Davis, Dept Comp Sci, 2063 Kemper Hall, Davis, CA 95616 USA
基金
美国国家科学基金会; 美国食品与农业研究所;
关键词
Automated crop monitoring; Non-destructive sensing for quality control; Visual anomaly detection; Data augmentation; Curriculum learning;
D O I
10.1016/j.compag.2024.109416
中图分类号
S [农业科学];
学科分类号
09 ;
摘要
Modern, automated quality control systems for speciality crops utilise computer vision together with a machine learning paradigm exploiting large datasets for learning efficient crop assessment components. To model anomalous visuals, data augmentation methods are often developed as a simple yet powerful tool for manipulating readily available normal samples. State-of-the-art augmentation methods embed arbitrary "structural"peculiarities in normal images to build a classifier of these artefacts (i.e., pretext task), enabling self-supervised representation learning of visual signals for anomaly detection (i.e., downstream task). In this paper, however, we argue that learning such structure-sensitive representations may be suboptimal for agricultural anomalies (e.g., unhealthy crops) that could be better recognised by a different type of visual element like "colour". To be specific, we propose Channel Randomisation (CH-Rand)-a novel data augmentation method that forces deep neural networks to learn effective encoding of "colour irregularities"under self-supervision whilst performing a pretext task to discriminate channel-randomised images. Extensive experiments are performed across various types of speciality crops (apples, strawberries, oranges, and bananas) to validate the informativeness of learnt representations in detecting anomalous instances. Our results demonstrate that CH-Rand's representations are significantly more reliable and robust, outperforming state-of-the-art methods (e.g., CutPaste) that learn structural representations by over 43% in Area Under the Precision-Recall Curve (AUC-PR), particularly for strawberries. Additional experiments suggest that adopting the L*a*b* * a * b * colour space and "curriculum"learning in the pretext task - gradually disregarding channel combinations for unrealistic outcomes - further improves downstream-task performance by 16% in AUC-PR. In particular, our experiments employ Riseholme-2021, , a novel speciality crop dataset consisting of 3.5K real strawberry images gathered in situ from the real farm, along with the Fresh & Stale public dataset. All our code and datasets are made publicly available online to ensure reproducibility and encourage further research in agricultural technologies.
引用
收藏
页数:15
相关论文
共 50 条
  • [31] Understanding the limitations of self-supervised learning for tabular anomaly detection
    Mai, Kimberly T.
    Davies, Toby
    Griffin, Lewis D.
    PATTERN ANALYSIS AND APPLICATIONS, 2024, 27 (02)
  • [32] SELF-SUPERVISED ACOUSTIC ANOMALY DETECTION VIA CONTRASTIVE LEARNING
    Hojjati, Hadi
    Armanfard, Narges
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 3253 - 3257
  • [33] Self-Supervised Video Forensics by Audio-Visual Anomaly Detection
    Feng, Chao
    Chen, Ziyang
    Owens, Andrew
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 10491 - 10503
  • [34] Audio-Visual Predictive Coding for Self-Supervised Visual Representation Learning
    Tellamekala, Mani Kumar
    Valstar, Michel
    Pound, Michael
    Giesbrecht, Timo
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 9912 - 9919
  • [35] Hyperspectral anomaly detection with self-supervised anomaly prior
    Liu, Yidan
    Jiang, Kai
    Xie, Weiying
    Zhang, Jiaqing
    Li, Yunsong
    Fang, Leyuan
    NEURAL NETWORKS, 2025, 187
  • [36] Boost Supervised Pretraining for Visual Transfer Learning: Implications of Self-Supervised Contrastive Representation Learning
    Sun, Jinghan
    Wei, Dong
    Ma, Kai
    Wang, Liansheng
    Zheng, Yefeng
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 2307 - 2315
  • [37] Self-Supervised Video Representation Learning by Video Incoherence Detection
    Cao, Haozhi
    Xu, Yuecong
    Mao, Kezhi
    Xie, Lihua
    Yin, Jianxiong
    See, Simon
    Xu, Qianwen
    Yang, Jianfei
    IEEE TRANSACTIONS ON CYBERNETICS, 2024, 54 (06) : 3810 - 3822
  • [38] Comparing Learning Methodologies for Self-Supervised Audio-Visual Representation Learning
    Terbouche, Hacene
    Schoneveld, Liam
    Benson, Oisin
    Othmani, Alice
    IEEE ACCESS, 2022, 10 : 41622 - 41638
  • [39] Whitening for Self-Supervised Representation Learning
    Ermolov, Aleksandr
    Siarohin, Aliaksandr
    Sangineto, Enver
    Sebe, Nicu
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [40] Self-Supervised Representation Learning for CAD
    Jones, Benjamin T.
    Hu, Michael
    Kodnongbua, Milin
    Kim, Vladimir G.
    Schulz, Adriana
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 21327 - 21336