Towards cosmological inference on unlabeled out-of-distribution HI observational data

被引:0
|
作者
Andrianomena, Sambatra [1 ,2 ]
Hassan, Sultan [2 ,3 ]
机构
[1] SARAO, Liesbeek House,River Pk Liesbeek Pkwy,Settlers Way, ZA-7705 Cape Town, South Africa
[2] Univ Western Cape, Dept Phys & Astron, ZA-7535 Cape Town, South Africa
[3] New York Univ, Ctr Cosmol & Particle Phys, Dept Phys, 726 Broadway, New York, NY 10003 USA
关键词
Large-scale structure of Universe; Methods:; numerical; statistical; Techniques: machine learning; ASTROPHYSICS;
D O I
10.1007/s10509-025-04405-y
中图分类号
P1 [天文学];
学科分类号
0704 ;
摘要
We present an approach that can be utilized in order to account for the covariate shift between two datasets of the same observable with different distributions. This helps improve the generalizability of a neural network model trained on in-distribution samples (IDs) when inferring cosmology at the field level on out-of-distribution samples (OODs) of unknown labels. We make use of HI maps from the two simulation suites in CAMELS, IllustrisTNG and SIMBA. We consider two different techniques, namely adversarial approach and optimal transport, to adapt a target network whose initial weightsare those of a source network pre-trained on a labeled dataset. Results show that after adaptation, salient features that are extracted by source and target encoders are well aligned in the embedding space. This indicates that the target encoder has learned the representations of the target domain via the adversarial training and optimal transport. Furthermore, in allscenarios considered in our analyses, the target encoder, which does not have access to any labels (ohm(m)) during adaptation phase, is able to retrieve the underlying ohm(m )from out-of-distribution maps to a great accuracy of R(2 )score >= 0.9, comparable to the performance of the source encoder trained in a supervised learning setup. We further test the viability of the techniques when only a few out-of-distribution instances are available for training and find that the target encoder still reasonably recovers the matter density. Our approach is critical in extracting information from upcoming large scale surveys.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Selecting Augmentation Methods for Domain Generalization and Out-of-Distribution Detection Using Unlabeled Data
    Kucuktas, Ulku Tuncer
    Uysal, Fatih
    Hardalac, Firat
    32ND IEEE SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE, SIU 2024, 2024,
  • [2] Exploiting Mixed Unlabeled Data for Detecting Samples of Seen and Unseen Out-of-Distribution Classes
    Sun, Yi-Xuan
    Wang, Wei
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 8386 - 8394
  • [3] FINE-GRAIN INFERENCE ON OUT-OF-DISTRIBUTION DATA WITH HIERARCHICAL CLASSIFICATION
    Linderman, Randolph
    Zhang, Jingyang
    Inkawhich, Nathan
    Li, Hai
    Chen, Yiran
    CONFERENCE ON LIFELONG LEARNING AGENTS, VOL 232, 2023, 232 : 162 - 183
  • [4] The Value of Out-of-Distribution Data
    De Silva, Ashwin
    Ramesh, Rahul
    Priebe, Carey E.
    Chaudhari, Pratik
    Vogelstein, Joshua T.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 202, 2023, 202
  • [5] Towards In-Distribution Compatible Out-of-Distribution Detection
    Wu, Boxi
    Jiang, Jie
    Ren, Haidong
    Du, Zifan
    Wang, Wenxiao
    Li, Zhifeng
    Cai, Deng
    He, Xiaofei
    Lin, Binbin
    Liu, Wei
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 9, 2023, : 10333 - 10341
  • [6] Towards a Theoretical Framework of Out-of-Distribution Generalization
    Ye, Haotian
    Xie, Chuanlong
    Cai, Tianle
    Li, Ruichen
    Li, Zhenguo
    Wang, Liwei
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [7] Causal Inference via Style Transfer for Out-of-distribution Generalisation
    Toan Nguyen
    Kien Do
    Duc Thanh Nguyen
    Bao Duong
    Thin Nguyen
    PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 1746 - 1757
  • [8] Causal inference for out-of-distribution recognition via sample balancing
    Wang, Yuqing
    Li, Xiangxian
    Liu, Yannan
    Cao, Xiao
    Meng, Xiangxu
    Meng, Lei
    CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY, 2024, 9 (05) : 1172 - 1184
  • [9] LEARNING WITH OUT-OF-DISTRIBUTION DATA FOR AUDIO CLASSIFICATION
    Iqbal, Turab
    Cao, Yin
    Kong, Qiuqiang
    Plumbley, Mark D.
    Wang, Wenwu
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 636 - 640
  • [10] Understanding Out-of-distribution:A Perspective of Data Dynamics
    Adila, Dyah
    Kang, Dongyeop
    WORKSHOP AT NEURIPS 2021, VOL 163, 2021, 163 : 1 - 8