Multi-modal Representation Learning for Social Post Location Inference

被引:0
|
作者
Dai, RuiTing [1 ]
Luo, Jiayi [1 ]
Luo, Xucheng [1 ]
Mo, Lisi [1 ]
Ma, Wanlun [2 ]
Zhou, Fan [1 ]
机构
[1] Univ Elect Sci & Technol China, Sch Informat & Software Engn, Chengdu 610054, Peoples R China
[2] Swinburne Univ Technol, Melbourne, Vic, Australia
来源
ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS | 2023年
关键词
Social geographic location; multi-modal social post dataset; multi-modal representation learning; multi-head attention mechanism; PREDICTION;
D O I
10.1109/ICC45041.2023.10279649
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
Inferring geographic locations via social posts is essential for many practical location-based applications such as product marketing, point-of-interest recommendation, and infector tracking for COVID-19. Unlike image-based location retrieval or social-post text embedding-based location inference, the combined effect of multi-modal information (i.e., post images, text, and hashtags) for social post positioning receives less attention. In this work, we collect real datasets of social posts with images, texts, and hashtags from Instagram and propose a novel Multi-modal Representation Learning Framework (MRLF) capable of fusing different modalities of social posts for location inference. MRLF integrates a multi-head attention mechanism to enhance location-salient information extraction while significantly improving location inference compared with single domain-based methods. To overcome the noisy user-generated textual content, we introduce a novel attention-based character-aware module that considers the relative dependencies between characters of social post texts and hashtags for flexible multimodel information fusion. The experimental results show that MRLF can make accurate location predictions and open a new door to understanding the multi-modal data of social posts for online inference tasks.
引用
收藏
页码:6331 / 6336
页数:6
相关论文
共 50 条
  • [31] Lightweight Multi-modal Representation Learning for RGB Salient Object Detection
    Yun Xiao
    Yameng Huang
    Chenglong Li
    Lei Liu
    Aiwu Zhou
    Jin Tang
    Cognitive Computation, 2023, 15 : 1868 - 1883
  • [32] Incomplete multi-modal representation learning for Alzheimer's disease diagnosis
    Liu, Yanbei
    Fan, Lianxi
    Zhang, Changqing
    Zhou, Tao
    Xiao, Zhitao
    Geng, Lei
    Shen, Dinggang
    MEDICAL IMAGE ANALYSIS, 2021, 69
  • [33] Multi-modal entity alignment based on joint knowledge representation learning
    Wang H.-Y.
    Lun B.
    Zhang X.-M.
    Sun X.-L.
    Kongzhi yu Juece/Control and Decision, 2021, 35 (12): : 2855 - 2864
  • [34] Deep Multi-modal Latent Representation Learning for Automated Dementia Diagnosis
    Zhou, Tao
    Liu, Mingxia
    Fu, Huazhu
    Wang, Jun
    Shen, Jianbing
    Shao, Ling
    Shen, Dinggang
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2019, PT IV, 2019, 11767 : 629 - 638
  • [35] CLMTR: a generic framework for contrastive multi-modal trajectory representation learning
    Liang, Anqi
    Yao, Bin
    Xie, Jiong
    Zheng, Wenli
    Shen, Yanyan
    Ge, Qiqi
    GEOINFORMATICA, 2024, : 233 - 253
  • [36] ONLINE INFERENCE WITH MULTI-MODAL LIKELIHOOD FUNCTIONS
    Gerber, Mathieu
    Heine, Kari
    ANNALS OF STATISTICS, 2021, 49 (06): : 3103 - 3126
  • [37] Unsupervised Multi-modal Learning
    Iqbal, Mohammed Shameer
    ADVANCES IN ARTIFICIAL INTELLIGENCE (AI 2015), 2015, 9091 : 343 - 346
  • [38] A Multi-Modal Active Learning Experience for Teaching Social Categorization
    Schwarzmueller, April
    TEACHING OF PSYCHOLOGY, 2011, 38 (03) : 158 - 161
  • [39] Learning Multi-modal Similarity
    McFee, Brian
    Lanckriet, Gert
    JOURNAL OF MACHINE LEARNING RESEARCH, 2011, 12 : 491 - 523
  • [40] DrFuse: Learning Disentangled Representation for Clinical Multi-Modal Fusion with Missing Modality and Modal Inconsistency
    Yao, Wenfang
    Yin, Kejing
    Cheung, William K.
    Liu, Jia
    Qin, Jing
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 15, 2024, : 16416 - 16424