SqueezeNeRF: Further factorized FastNeRF for memory-efficient inference

被引:12
|
作者
Wadhwani, Krishna [1 ]
Kojima, Tamaki [1 ]
机构
[1] Sony Grp Corp, Nihonbashi, Tokyo, Japan
关键词
D O I
10.1109/CVPRW56347.2022.00307
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Neural Radiance Fields (NeRF) has emerged as the state-of-the-art method for novel view generation of complex scenes, but is very slow during inference. Recently, there have been multiple works on speeding up NeRF inference, but the state of the art methods for real-time NeRF inference rely on caching the neural network output, which occupies several giga-bytes of disk space that limits their real-world applicability. As caching the neural network of original NeRF network is not feasible, Garbin et al. proposed "FastNeRF" which factorizes the problem into 2 subnetworks - one which depends only on the 3D coordinate of a sample point and one which depends only on the 2D camera viewing direction. Although this factorization enables them to reduce the cache size and perform inference at over 200 frames per second, the memory overhead is still substantial. In this work, we propose SqueezeNeRF, which is more than 60 times memory-efficient than the sparse cache of FastNeRF and is still able to render at more than 190 frames per second on a high spec GPU during inference.
引用
收藏
页码:2716 / 2724
页数:9
相关论文
共 50 条
  • [21] Memory-Efficient and Secure DNN Inference on TrustZone-enabled Consumer IoT Devices
    Xie, Xueshuo
    Wang, Haoxu
    Jian, Zhaolong
    Li, Tao
    Wang, Wei
    Xu, Zhiwei
    Wang, Guiling
    IEEE INFOCOM 2024-IEEE CONFERENCE ON COMPUTER COMMUNICATIONS, 2024, : 2009 - 2018
  • [22] Block Convolution: Toward Memory-Efficient Inference of Large-Scale CNNs on FPGA
    Li, Gang
    Liu, Zejian
    Li, Fanrong
    Cheng, Jian
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2022, 41 (05) : 1436 - 1447
  • [23] Memory-Efficient Adaptive Optimization
    Anil, Rohan
    Gupta, Vineet
    Koren, Tomer
    Singer, Yoram
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [24] Activation Sequence Caching: High-Throughput and Memory-Efficient Generative Inference with a Single GPU
    Kim, Sowoong
    Sim, Eunyeong
    Shin, Youngsam
    Cho, YeonGon
    Baek, Woongki
    PROCEEDINGS OF THE 2024 THE INTERNATIONAL CONFERENCE ON PARALLEL ARCHITECTURES AND COMPILATION TECHNIQUES, PACT 2024, 2024, : 78 - 90
  • [25] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning
    Lin, Ji
    Chen, Wei-Ming
    Cai, Han
    Gan, Chuang
    Han, Song
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [26] Smart-DNN plus : A Memory-efficient Neural Networks Compression Framework for the Model Inference
    Wu, Donglei
    Yang, Weihao
    Zou, Xiangyu
    Xia, Wen
    Li, Shiyi
    Hu, Zhenbo
    Zhang, Weizhe
    Fang, Binxing
    ACM TRANSACTIONS ON ARCHITECTURE AND CODE OPTIMIZATION, 2023, 20 (04)
  • [27] A memory-efficient strategy for exploring the web
    Castillo, Carlos
    Nelli, Alberto
    Panconesi, Alessandro
    2006 IEEE/WIC/ACM INTERNATIONAL CONFERENCE ON WEB INTELLIGENCE, (WI 2006 MAIN CONFERENCE PROCEEDINGS), 2006, : 680 - +
  • [28] Memory-efficient Parallel Tensor Decompositions
    Baskaran, Muthu
    Henretty, Tom
    Pradelle, Benoit
    Langston, M. Harper
    Bruns-Smith, David
    Ezick, James
    Lethin, Richard
    2017 IEEE HIGH PERFORMANCE EXTREME COMPUTING CONFERENCE (HPEC), 2017,
  • [29] Toward memory-efficient linear solvers
    Baker, A
    Dennis, J
    Jessup, ER
    HIGH PERFORMANCE COMPUTING FOR COMPUTATIONAL SCIENCE - VECPAR 2002, 2003, 2565 : 315 - 327
  • [30] Memory-Efficient Assembly Using Flye
    Freire, Borja
    Ladra, Susana
    Parama, Jose R.
    IEEE-ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS, 2022, 19 (06) : 3564 - 3577