Causal invariant geographic network representations with feature and structural distribution shifts

被引:0
|
作者
Wang, Yuhan [1 ]
He, Silu [1 ]
Luo, Qinyao [1 ]
Yuan, Hongyuan [1 ]
Zhao, Ling [1 ]
Zhu, Jiawei [2 ]
Li, Haifeng [1 ]
机构
[1] Cent South Univ, Dept Geog Informat Syst, Changsha 410000, Hunan, Peoples R China
[2] Cent South Univ, Sch Architecture & Art, Changsha 410083, Hunan, Peoples R China
关键词
Geographic network representation learning; Out-of-distribution generalisation; Casual inference;
D O I
10.1016/j.future.2025.107814
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Relationships between geographic entities, including human-land and human-people relationships, can be naturally modelled by graph structures, and geographic network representation is an important theoretical issue. The existing methods learn geographic network representations through deep graph neural networks (GNNs) based on the i.i.d. assumption. However, the spatial heterogeneity and temporal dynamics of geographic data make the out-of-distribution (OOD) generalisation problem particularly salient. We classify geographic network representations into invariant representations that always stabilise the predicted labels under distribution shifts and background representations that vary with different distributions. The latter are particularly sensitive to distribution shifts (feature and structural shifts) between testing and training data and are the main causes of the out-of-distribution generalisation (OOD) problem. Spurious correlations are present between invariant and background representations due to selection biases/environmental effects, resulting in the model extremes being more likely to learn background representations. The existing approaches focus on background representation changes that are determined by shifts in the feature distributions of nodes in the training and test data while ignoring changes in the proportional distributions of heterogeneous and homogeneous neighbour nodes, which we refer to as structural distribution shifts. We propose a feature-structure mixed invariant representation learning (FSM-IRL) model that accounts for both feature distribution shifts and structural distribution shifts. To address structural distribution shifts, we introduce a sampling method based on causal attention, encouraging the model to identify nodes possessing strong causal relationships with labels or nodes that are more similar to the target node. This approach significantly enhances the invariance of the representations between the source and target domains while reducing the dependence on background representations that arise by chance or in specific patterns. Inspired by the Hilbert-Schmidt independence criterion, we implement a reweighting strategy to maximise the orthogonality of the node representations, thereby mitigating the spurious correlations among the node representations and suppressing the learning of background representations. In addition, we construct an educational-level geographic network dataset under out-of-distribution (OOD) conditions. Our experiments demonstrate that FSM-IRL exhibits strong learning capabilities on both geographic and social network datasets in OOD scenarios.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Invariant and Transportable Representations for Anti-Causal Domain Shifts
    Jiang, Yibo
    Veitch, Victor
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [2] Invariant Anomaly Detection under Distribution Shifts: A Causal Perspective
    Carvalho, Joao B. S.
    Zhang, Mengtao
    Geyer, Robin
    Cotrini, Carlos
    Buhmann, Joachim M.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [3] Diagnosing and Repairing Feature Representations under Distribution Shifts
    Lourenco, Ines
    Bobu, Andreea
    Rojas, Cristian R.
    Wahlberg, Bo
    2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 3638 - 3645
  • [4] Estimating Generalization under Distribution Shifts via Domain-Invariant Representations
    Chuang, Ching-Yao
    Torralba, Antonio
    Jegelka, Stefanie
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [5] Estimating Generalization under Distribution Shifts via Domain-Invariant Representations
    Chuang, Ching-Yao
    Torralba, Antonio
    Jegelka, Stefanie
    25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019), 2019,
  • [6] Neuroevolutionary Feature Representations for Causal Inference
    Burkhart, Michael C.
    Ruiz, Gabriel
    COMPUTATIONAL SCIENCE, ICCS 2022, PT II, 2022, : 3 - 10
  • [7] Robust Reconfigurable Intelligent Surfaces via Invariant Risk and Causal Representations
    Samarakoon, Sumudu
    Park, Jihong
    Bennis, Mehdi
    SPAWC 2021: 2021 IEEE 22ND INTERNATIONAL WORKSHOP ON SIGNAL PROCESSING ADVANCES IN WIRELESS COMMUNICATIONS (IEEE SPAWC 2021), 2020, : 301 - 305
  • [8] Out-of-distribution Generalization with Causal Invariant Transformations
    Wang, Ruoyu
    Yi, Mingyang
    Chen, Zhitang
    Zhu, Shengyu
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 375 - 385
  • [9] Invariant Feature Learning Based on Causal Inference from Heterogeneous Environments
    Su, Hang
    Wang, Wei
    MATHEMATICS, 2024, 12 (05)
  • [10] CLASSIFICATION OF INVARIANT IMAGE REPRESENTATIONS USING A NEURAL NETWORK
    KHOTANZAD, A
    LU, JH
    IEEE TRANSACTIONS ON ACOUSTICS SPEECH AND SIGNAL PROCESSING, 1990, 38 (06): : 1028 - 1038