OmniCity: Omnipotent City Understanding with Multi-level and Multi-view Images

被引:6
|
作者
Li, Weijia [1 ]
Lai, Yawen [2 ]
Xu, Linning [3 ]
Xiangli, Yuanbo [3 ]
Yu, Jinhua [1 ]
He, Conghui [2 ,4 ]
Xia, Gui-Song [5 ]
Lin, Dahua [3 ,4 ]
机构
[1] Sun Yat Sen Univ, Guangzhou, Peoples R China
[2] SenseTime Res, Hong Kong, Peoples R China
[3] Chinese Univ Hong Kong, Hong Kong, Peoples R China
[4] Shanghai Artificial Intelligence Lab, Shanghai, Peoples R China
[5] Wuhan Univ, Wuhan, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/CVPR52729.2023.01669
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents OmniCity, a new dataset for omnipotent city understanding from multi-level and multi-view images. More precisely, OmniCity contains multi-view satellite images as well as street-level panorama and monoview images, constituting over 100K pixel-wise annotated images that are well-aligned and collected from 25K geolocations in New York City. To alleviate the substantial pixel-wise annotation efforts, we propose an efficient street-view image annotation pipeline that leverages the existing label maps of satellite view and the transformation relations between different views (satellite, panorama, and mono-view). With the new OmniCity dataset, we provide benchmarks for a variety of tasks including building footprint extraction, height estimation, and building plane/instance/fine-grained segmentation. Compared with existing multi-level and multi-view benchmarks, OmniCity contains a larger number of images with richer annotation types and more views, provides more benchmark results of state-of-the-art models, and introduces a new task for fine-grained building instance segmentation on street-level panorama images. Moreover, OmniCity provides new problem settings for existing tasks, such as cross-view image matching, synthesis, segmentation, detection, etc., and facilitates the developing of new methods for large-scale city understanding, reconstruction, and simulation. The OmniCity dataset as well as the benchmarks will be released at https://city-super.github.io/omnicity/.
引用
收藏
页码:17397 / 17407
页数:11
相关论文
共 50 条
  • [41] VRML animation from multi-view images
    Iwadate, Y
    Katayama, M
    Tomiyama, K
    Imaizumi, H
    IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, VOL I AND II, PROCEEDINGS, 2002, : 881 - 884
  • [42] Generating Virtual Images for Multi-view Video
    ZHANG Lijun
    LI Yue
    ZHU Qiuyu
    LI Mingqi
    Chinese Journal of Electronics, 2017, 26 (04) : 810 - 813
  • [43] Coding of multi-view images for immersive viewing
    Tong, X
    Gray, RM
    2000 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, PROCEEDINGS, VOLS I-VI, 2000, : 1879 - 1882
  • [44] Multi-view semantic understanding for visual dialog
    Jiang, Tianling
    Zhang, Zefan
    Li, Xin
    Ji, Yi
    Liu, Chunping
    KNOWLEDGE-BASED SYSTEMS, 2023, 268
  • [45] Learning multi-level topology representation for multi-view clustering with deep non-negative matrix factorization
    Dou, Zengfa
    Peng, Nian
    Hou, Weiming
    Xie, Xianghua
    Ma, Xiaoke
    NEURAL NETWORKS, 2025, 182
  • [46] Walking Imagery Evaluation in Brain Computer Interfaces via a Multi-view Multi-level Deep Polynomial Network
    Lei, Baiying
    Liu, Xiaolu
    Liang, Shuang
    Hang, Wenlong
    Wang, Qiong
    Choi, Kup-Sze
    Qin, Jing
    IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING, 2019, 27 (03) : 497 - 506
  • [47] Adaptive structural-guided multi-level representation learning with graph contrastive for incomplete multi-view clustering
    Wang, Haiyue
    Zhang, Wensheng
    Wang, Quan
    Ma, Xiaoke
    INFORMATION FUSION, 2025, 119
  • [48] Multi-level multi-view network based on structural contrastive learning for scRNA-seq data clustering
    Shu, Zhenqiu
    Xia, Min
    Tan, Kaiwen
    Zhang, Yongbing
    Yu, Zhengtao
    BRIEFINGS IN BIOINFORMATICS, 2024, 25 (06)
  • [49] On Multi-Level Thinking and Scientific Understanding
    Michael Edgeworth McINTYRE
    AdvancesinAtmosphericSciences, 2017, 34 (10) : 1150 - 1158
  • [50] On multi-level thinking and scientific understanding
    Michael Edgeworth McIntyre
    Advances in Atmospheric Sciences, 2017, 34 : 1150 - 1158