Automatic report generation based on multi-modal information

被引:0
|
作者
Jing Zhang
Xiaoxue Li
Weizhi Nie
Yuting Su
机构
[1] Tianjin University,School of Electronics Information Engineering
来源
关键词
News event detection; Multi-modal; Report generation;
D O I
暂无
中图分类号
学科分类号
摘要
In this paper, we propose a new framework which can utilize multi-modal social media information to automatically generate related reports for users or government. First, we utilize DBSCAN (Density Based Spatial Clustering of Applications with Noise) to detect events in official news websites. Then, some unofficial information details are extracted from social network platforms (Foursquare, Twitter, YouTube), which will be leveraged to enhance the official report in order to excavate some latent and useful information. In this process, we applied some classic textual processing methods and computer vision technologies to reduce the noise information uploaded by user generated contents (UGCs). Then, we applied LSTM-CNN model to generate the related image caption and successfully convert visual information to textual information. Finally, we extracted some latent topics using graph cluster method to generate the final report. To demonstrate the effectiveness of our framework, we got a large of multi-source event dataset from official news websites and Twitter. Finally, the user study demonstrates the practicability of our approach.
引用
收藏
页码:12005 / 12015
页数:10
相关论文
共 50 条
  • [31] A data-based framework for automatic road network generation of multi-modal transport micro-simulation
    Zhang, Qi
    Wang, Yukai
    Yin, Ruyang
    Cheng, Wenyu
    Wan, Jian
    Wu, Lan
    ELECTRONIC RESEARCH ARCHIVE, 2022, 31 (01): : 190 - 206
  • [32] Multi-modal transformer architecture for medical image analysis and automated report generation
    Raminedi, Santhosh
    Shridevi, S.
    Won, Daehan
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [33] A Multi-Modal Chinese Poetry Generation Model
    Liu, Dayiheng
    Guo, Quan
    Li, Wubo
    2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,
  • [34] Multi-modal Sarcasm Generation: Dataset and Solution
    Zhao, Wenye
    Huang, Qingbao
    Xu, Dongsheng
    Zhao, Peizhi
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 5601 - 5613
  • [35] Meme Generation with Multi-modal Input and Planning
    Ranjan, Ashutosh
    Srivastava, Vivek
    Khatri, Jyotsana
    Bhat, Savita
    Karande, Shirish
    PROCEEDINGS OF THE 2ND INTERNATIONAL WORKSHOP ON DEEP MULTIMODAL GENERATION AND RETRIEVAL, MMGR 2024, 2024, : 21 - +
  • [36] Multi-modal information retrieval using FINT
    van Zaanen, M
    de Croon, G
    MULTILINGUAL INFORMATION ACCESS FOR TEXT, SPEECH AND IMAGES, 2005, 3491 : 728 - +
  • [37] Multi-modal Recommendation System with Auxiliary Information
    Muthivhi, Mufhumudzi
    van Zyl, Terence
    Wang, Hairong
    ARTIFICIAL INTELLIGENCE RESEARCH, SACAIR 2022, 2022, 1734 : 108 - 122
  • [38] Emotion Recognition from Multi-Modal Information
    Wu, Chung-Hsien
    Lin, Jen-Chun
    Wei, Wen-Li
    Cheng, Kuan-Chun
    2013 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA), 2013,
  • [39] Towards Multi-Modal Conversational Information Seeking
    Deldjoo, Yashar
    Trippas, Johanne R.
    Zamani, Hamed
    SIGIR '21 - PROCEEDINGS OF THE 44TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2021, : 1577 - 1587
  • [40] Multi-modal Information Integration for Document Retrieval
    Hassan, Ehtesham
    Chaudhury, Santanu
    Gopal, M.
    2013 12TH INTERNATIONAL CONFERENCE ON DOCUMENT ANALYSIS AND RECOGNITION (ICDAR), 2013, : 1200 - 1204