Active Object Detection with Knowledge Aggregation and Distillation from Large Models

被引:2
|
作者
Yang, Dejie [1 ]
Liu, Yang [1 ]
机构
[1] Peking Univ, Wangxuan Inst Comp Technol, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/CVPR52733.2024.01573
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Accurately detecting active objects undergoing state changes is essential for comprehending human interactions and facilitating decision-making. The existing methods for active object detection (AOD) primarily rely on visual appearance of the objects within input, such as changes in size, shape and relationship with hands. However, these visual changes can be subtle, posing challenges, particularly in scenarios with multiple distracting no-change instances of the same category. We observe that the state changes are often the result of an interaction being performed upon the object, thus propose to use informed priors about object related plausible interactions (including semantics and visual appearance) to provide more reliable cues for AOD. Specifically, we propose a knowledge aggregation procedure to integrate the aforementioned informed priors into oracle queries within the teacher decoder, offering more object affordance commonsense to locate the active object. To streamline the inference process and reduce extra knowledge inputs, we propose a knowledge distillation approach that encourages the student decoder to mimic the detection capabilities of the teacher decoder using the oracle query by replicating its predictions and attention. Our proposed framework achieves state-of-the-art performance on four datasets, namely Ego4D, Epic-Kitchens, MECCANO, and 100DOH, which demonstrates the effectiveness of our approach in improving AOD. The code and models are available at https://github.com/idejie/KAD.git.
引用
收藏
页码:16624 / 16633
页数:10
相关论文
共 50 条
  • [31] ROBUST AND ACCURATE OBJECT DETECTION VIA SELF-KNOWLEDGE DISTILLATION
    Xu, Weipeng
    Chu, Pengzhi
    Xie, Renhao
    Xiao, Xiongziyan
    Huang, Hongcheng
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 91 - 95
  • [32] Revisiting knowledge distillation for light-weight visual object detection
    Gao, Tianze
    Gao, Yunfeng
    Li, Yu
    Qin, Peiyuan
    TRANSACTIONS OF THE INSTITUTE OF MEASUREMENT AND CONTROL, 2021, 43 (13) : 2888 - 2898
  • [33] GAN-Knowledge Distillation for One-Stage Object Detection
    Wang, Wanwei
    Hong, Wei
    Wang, Feng
    Yu, Jinke
    IEEE ACCESS, 2020, 8 : 60719 - 60727
  • [34] Knowledge Distillation in Object Detection for Resource-Constrained Edge Computing
    Setyanto, Arief
    Sasongko, Theopilus Bayu
    Fikri, Muhammad Ainul
    Ariatmanto, Dhani
    Agastya, I. Made Artha
    Rachmanto, Rakandhiya Daanii
    Ardana, Affan
    Kim, In Kee
    IEEE ACCESS, 2025, 13 : 18200 - 18214
  • [35] Towards Efficient 3D Object Detection with Knowledge Distillation
    Yang, Jihan
    Shi, Shaoshuai
    Ding, Runyu
    Wang, Zhe
    Qi, Xiaojuan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [36] Multidomain Object Detection Framework Using Feature Domain Knowledge Distillation
    Jaw, Da-Wei
    Huang, Shih-Chia
    Lu, Zhi-Hui
    Fung, Benjamin C. M.
    Kuo, Sy-Yen
    IEEE TRANSACTIONS ON CYBERNETICS, 2024, 54 (08) : 4643 - 4651
  • [37] Scalability of knowledge distillation in incremental deep learning for fast object detection
    Yuwono, Elizabeth Irenne
    Tjondonegoro, Dian
    Sorwar, Golam
    Alaei, Alireza
    APPLIED SOFT COMPUTING, 2022, 129
  • [38] Active Object Detection with Epistemic Uncertainty and Hierarchical Information Aggregation
    Park, Younghyun
    Kim, Soyeong
    Choi, Wonjeong
    Han, Dong-Jun
    Moon, Jaekyun
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2022, 2022, : 2711 - 2715
  • [39] Towards Interpreting Vulnerability of Object Detection Models via Adversarial Distillation
    Zhang, Yaoyuan
    Tan, Yu-an
    Lu, Mingfeng
    Liu, Lu
    Zhang, Quanxing
    Li, Yuanzhang
    Wang, Dianxin
    APPLIED CRYPTOGRAPHY AND NETWORK SECURITY WORKSHOPS, ACNS 2022, 2022, 13285 : 53 - 65
  • [40] Towards interpreting vulnerability of object detection models via adversarial distillation
    Zhang, Yaoyuan
    Tan, Yu-an
    Lu, Mingfeng
    Liu, Lu
    Wang, Dianxin
    Zhang, Quanxing
    Li, Yuanzhang
    JOURNAL OF INFORMATION SECURITY AND APPLICATIONS, 2023, 73