A Word Embedding Model Learned from Political Tweets

被引:0
|
作者
Alnajran, Noufa N. [1 ]
Crockett, Keeley A. [1 ]
McLean, David [1 ]
Latham, Annabel [1 ]
机构
[1] Manchester Metropolitan Univ, Dept Comp Math & Digital Technol, Manchester, Lancs, England
关键词
Word Embedding; Language Modelling; Deep Learning; Social Network Analysis; Twitter Analysis;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Distributed word representations have recently contributed to significant improvements in many natural language processing (NLP) tasks. Distributional semantics have become amongst the important trends in machine learning (ML) applications. Word embeddings are distributed representations of words that learn semantic relationships from a large corpus of text. In the social context, the distributed representation of a word is likely to be different from general text word embeddings. This is relatively due to the unique lexical semantic features and morphological structure of social media text such as tweets, which implies different word vector representations. In this paper, we collect and present a political social dataset that consists of over four million English tweets. An artificial neural network (NN) is trained to learn word co -occurrence and generate word vectors from the political corpus of tweets. The model is 136MB and includes word representations for a vocabulary of over 86K unique words and phrases. The learned model shall contribute to the success of many ML and NLP applications in microblogging Social Network Analysis (OSN), such as semantic similarity and cluster analysis tasks.
引用
收藏
页码:177 / 183
页数:7
相关论文
共 50 条
  • [31] An Enhanced Neural Word Embedding Model for Transfer Learning
    Kowsher, Md
    Sobuj, Md Shohanur Islam
    Shahriar, Md Fahim
    Prottasha, Nusrat Jahan
    Arefin, Mohammad Shamsul
    Dhar, Pranab Kumar
    Koshiba, Takeshi
    APPLIED SCIENCES-BASEL, 2022, 12 (06):
  • [32] Evaluation of Gender Bias in Amharic Word Embedding Model
    Zenebe, Beimnet
    Gizaw, Solomon
    Abgaz, Yalemisew
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS, PT II, NLDB 2024, 2024, 14763 : 172 - 182
  • [33] A Word Embedding Model for Analyzing Patterns and Their Distributional Semantics
    Feng, Rui
    Yang, Congcong
    Qu, Yunhua
    JOURNAL OF QUANTITATIVE LINGUISTICS, 2022, 29 (01) : 80 - 105
  • [34] Content Linking for UGC based on Word Embedding Model
    Gao, Zhiqiao
    Li, Lei
    Mao, Liyuan
    He, Dezhu
    Xue, Chao
    Proceedings of the 11th EAI International Conference on Heterogeneous Networking for Quality, Reliability, Security and Robustness, 2015, : 149 - 154
  • [35] A Character-Enhanced Chinese Word Embedding Model
    Yang, Gang
    Xu, Hongzhe
    He, Tianhao
    Cai, Zaishang
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [36] A ⟨word, part of speech⟩ embedding model for text classification
    Liu, Wenfeng
    Liu, Peiyu
    Yang, Yuzhen
    Yi, Jing
    Zhu, Zhenfang
    EXPERT SYSTEMS, 2019, 36 (06)
  • [37] The Learned Word
    Taylor, Kathleen
    PHI DELTA KAPPAN, 2009, 91 (01) : 7 - 7
  • [38] The Learned Word
    Taylor, Kathleen
    PHI DELTA KAPPAN, 2009, 91 (04) : 7 - 7
  • [39] The Learned Word
    Taylor, Kathleen
    PHI DELTA KAPPAN, 2010, 92 (02) : 7 - 7
  • [40] The Learned Word
    Taylor, Kathleen
    PHI DELTA KAPPAN, 2011, 92 (06) : 7 - 7