SqueezeNeRF: Further factorized FastNeRF for memory-efficient inference

被引:12
|
作者
Wadhwani, Krishna [1 ]
Kojima, Tamaki [1 ]
机构
[1] Sony Grp Corp, Nihonbashi, Tokyo, Japan
关键词
D O I
10.1109/CVPRW56347.2022.00307
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Neural Radiance Fields (NeRF) has emerged as the state-of-the-art method for novel view generation of complex scenes, but is very slow during inference. Recently, there have been multiple works on speeding up NeRF inference, but the state of the art methods for real-time NeRF inference rely on caching the neural network output, which occupies several giga-bytes of disk space that limits their real-world applicability. As caching the neural network of original NeRF network is not feasible, Garbin et al. proposed "FastNeRF" which factorizes the problem into 2 subnetworks - one which depends only on the 3D coordinate of a sample point and one which depends only on the 2D camera viewing direction. Although this factorization enables them to reduce the cache size and perform inference at over 200 frames per second, the memory overhead is still substantial. In this work, we propose SqueezeNeRF, which is more than 60 times memory-efficient than the sparse cache of FastNeRF and is still able to render at more than 190 frames per second on a high spec GPU during inference.
引用
收藏
页码:2716 / 2724
页数:9
相关论文
共 50 条
  • [1] Memory-Efficient Dataflow Inference for Deep CNNs on FPGA
    Petrica, Lucian
    Alonso, Tobias
    Kroes, Mairin
    Fraser, Nicholas
    Cotofana, Sorin
    Blott, Michaela
    2020 INTERNATIONAL CONFERENCE ON FIELD-PROGRAMMABLE TECHNOLOGY (ICFPT 2020), 2020, : 48 - 55
  • [2] Occamy: Memory-efficient GPU Compiler for DNN Inference
    Lee, Jaeho
    Jeong, Shinnung
    Song, Seungbin
    Kim, Kunwoo
    Choi, Heelim
    Kim, Youngsok
    Kim, Hanjun
    2023 60TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC, 2023,
  • [3] TETRIS: Memory-efficient Serverless Inference through Tensor Sharing
    Li, Jie
    Zhao, Laiping
    Yang, Yanan
    Zhan, Kunlin
    Li, Keqiu
    PROCEEDINGS OF THE 2022 USENIX ANNUAL TECHNICAL CONFERENCE, 2022, : 473 - 488
  • [4] Memory-Efficient Deep Learning Inference in Trusted Execution Environments
    Truong, Jean-Baptiste
    Gallagher, William
    Guo, Tian
    Walls, Robert J.
    2021 IEEE INTERNATIONAL CONFERENCE ON CLOUD ENGINEERING, IC2E 2021, 2021, : 161 - 167
  • [5] StreamNet: Memory-Efficient Streaming Tiny Deep Learning Inference on the Microcontroller
    Zheng, Hong-Sheng
    Hsu, Chen-Fong
    Liu, Yu-Yuan
    Yeh, Tsung Tai
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [6] Performance Trade-offs in Weight Quantization for Memory-Efficient Inference
    Tostado, Pablo M.
    Pedroni, Bruno U.
    Cauwenberghs, Gert
    2019 IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS 2019), 2019, : 246 - 250
  • [7] Evolutionary Bin Packing for Memory-Efficient Dataflow Inference Acceleration on FPGA
    Kroes, Mairin
    Petrica, Lucian
    Cotofana, Sorin
    Blott, Michaela
    GECCO'20: PROCEEDINGS OF THE 2020 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, 2020, : 1125 - 1133
  • [8] A Memory-Efficient Edge Inference Accelerator with XOR-based Model Compression
    Lee, Hyunseung
    Hong, Jihoon
    Kim, Soosung
    Lee, Seung Yul
    Lee, Jae W.
    2023 60TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC, 2023,
  • [9] PENETRALIUM: Privacy-preserving and memory-efficient neural network inference at the edge
    Yang, Mengda
    Yi, Wenzhe
    Wang, Juan
    Hu, Hongxin
    Xu, Xiaoyang
    Li, Ziang
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2024, 156 : 30 - 41
  • [10] Buffer Sizes Reduction for Memory-efficient CNN Inference on Mobile and Embedded Devices
    Minakova, Svetlana
    Stefanov, Todor
    2020 23RD EUROMICRO CONFERENCE ON DIGITAL SYSTEM DESIGN (DSD 2020), 2020, : 133 - 140