Towards Visually Explaining Variational Autoencoders

被引:111
|
作者
Liu, Wenqian [1 ]
Li, Runze [2 ]
Zheng, Meng [3 ]
Karanam, Srikrishna [4 ]
Wu, Ziyan [4 ]
Bhanu, Bir [2 ]
Radke, Richard J. [3 ]
Camps, Octavia [1 ]
机构
[1] Northeastern Univ, Boston, MA 02115 USA
[2] Univ Calif Riverside, Riverside, CA 92521 USA
[3] Rensselaer Polytech Inst, Troy, NY USA
[4] United Imaging Intelligence, Cambridge, MA USA
关键词
D O I
10.1109/CVPR42600.2020.00867
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent advances in convolutional neural network (CNN) model interpretability have led to impressive progress in visualizing and understanding model predictions. In particular, gradient-based visual attention methods have driven much recent effort in using visual attention maps as a means for visual explanations. A key problem, however, is these methods are designed for classification and categorization tasks, and their extension to explaining generative models, e.g., variational autoencoders (VAE) is not trivial. In this work, we take a step towards bridging this crucial gap, proposing the first technique to visually explain VAEs by means of gradient-based attention. We present methods to generate visual attention from the learned latent space, and also demonstrate such attention explanations serve more than just explaining VAE predictions. We show how these attention maps can be used to localize anomalies in images, demonstrating state-of-the-art performance on the MVTec-AD dataset. We also show how they can be infused into model training, helping bootstrap the VAE into learning improved latent space disentanglement, demonstrated on the Dsprites dataset.
引用
收藏
页码:8639 / 8648
页数:10
相关论文
共 50 条
  • [1] Towards Evaluating the Representation Learned by Variational AutoEncoders
    Ueda, Tatsuya
    Vargas, Danilo Vasconcellos
    2021 60TH ANNUAL CONFERENCE OF THE SOCIETY OF INSTRUMENT AND CONTROL ENGINEERS OF JAPAN (SICE), 2021, : 591 - 594
  • [2] Towards a Theoretical Understanding of the Robustness of Variational Autoencoders
    Camuto, Alexander
    Willetts, Matthew
    Roberts, Stephen
    Holmes, Chris
    Rainforth, Tom
    24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130
  • [3] Towards Visually Explaining Video Understanding Networks with Perturbation
    Li, Zhenqiang
    Wang, Weimin
    Li, Zuoyue
    Huang, Yifei
    Sato, Yoichi
    2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2021), 2021, : 1119 - 1128
  • [4] Towards Data-Driven Volatility Modeling with Variational Autoencoders
    Dierckx, Thomas
    Davis, Jesse
    Schoutens, Wim
    MACHINE LEARNING AND PRINCIPLES AND PRACTICE OF KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2022, PT II, 2023, 1753 : 97 - 111
  • [5] GraphVAE: Towards Generation of Small Graphs Using Variational Autoencoders
    Simonovsky, Martin
    Komodakis, Nikos
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2018, PT I, 2018, 11139 : 412 - 422
  • [6] Towards learning transferable embeddings for protein conformations using Variational Autoencoders
    Albu, Alexandra-Ioana
    KNOWLEDGE-BASED AND INTELLIGENT INFORMATION & ENGINEERING SYSTEMS (KSE 2021), 2021, 192 : 10 - 19
  • [7] Mixture variational autoencoders
    Jiang, Shuoran
    Chen, Yarui
    Yang, Jucheng
    Zhang, Chuanlei
    Zhao, Tingting
    PATTERN RECOGNITION LETTERS, 2019, 128 : 263 - 269
  • [8] An Introduction to Variational Autoencoders
    Kingma, Diederik P.
    Welling, Max
    FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2019, 12 (04): : 4 - 89
  • [9] Subitizing with Variational Autoencoders
    Wever, Rijnder
    Runia, Tom F. H.
    COMPUTER VISION - ECCV 2018 WORKSHOPS, PT III, 2019, 11131 : 617 - 627
  • [10] Mixtures of Variational Autoencoders
    Ye, Fei
    Bors, Adrian G.
    2020 TENTH INTERNATIONAL CONFERENCE ON IMAGE PROCESSING THEORY, TOOLS AND APPLICATIONS (IPTA), 2020,