Learning multi-prototype word embedding from single-prototype word embedding with integrated knowledge

被引:9
|
作者
Yang, Xuefeng [1 ]
Mao, Kezhi [1 ]
机构
[1] Nanyang Technol Univ, Sch Elect & Elect Engn, Nanyang Ave, Singapore 639798, Singapore
关键词
Multi-prototype word embedding; Distributional semantic model; Fine tuning; Semantic similarity; SEMANTIC SIMILARITY;
D O I
10.1016/j.eswa.2016.03.013
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Distributional semantic models (DSM) or word embeddings are widely used in prediction of semantic similarity and relatedness. However, context aware similarity and relatedness prediction is still a challenging issue because most DSM models or word embeddings use one vector per word without considering polysemy and homonym. In this paper, we propose a supervised fine tuning framework to transform the existing single-prototype word embeddings into multi-prototype word embeddings based on lexical semantic resources. As a post-processing step, the proposed framework is compatible with any sense inventory and any word embedding. To test the proposed learning framework, both intrinsic and extrinsic evaluations are conducted. Experiments results of 3 tasks with 8 datasets show that the multi-prototype word representations learned by the proposed framework outperform single-prototype word representations. (C) 2016 Elsevier Ltd. All rights reserved.
引用
收藏
页码:291 / 299
页数:9
相关论文
共 50 条
  • [1] Bridging Text and Knowledge by Learning Multi-Prototype Entity Mention Embedding
    Cao, Yixin
    Huang, Lifu
    Ji, Heng
    Chen, Xu
    Li, Juanzi
    PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, : 1623 - 1633
  • [2] Multi-prototype Chinese Character Embedding
    Lu, Yanan
    Zhang, Yue
    Ji, Donghong
    LREC 2016 - TENTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2016, : 855 - 859
  • [3] Modeling multi-prototype Chinese word representation learning for word similarity
    Fulian Yin
    Yanyan Wang
    Jianbo Liu
    Marco Tosato
    Complex & Intelligent Systems, 2021, 7 : 2977 - 2990
  • [4] Modeling multi-prototype Chinese word representation learning for word similarity
    Yin, Fulian
    Wang, Yanyan
    Liu, Jianbo
    Tosato, Marco
    COMPLEX & INTELLIGENT SYSTEMS, 2021, 7 (06) : 2977 - 2990
  • [5] Convolution-deconvolution word embedding: An end-to-end multi-prototype fusion embedding method for natural language processing
    Shuang, Kai
    Zhang, Zhixuan
    Loo, Jonathan
    Su, Sen
    INFORMATION FUSION, 2020, 53 : 112 - 122
  • [6] Incorporating Domain Knowledge in Learning Word Embedding
    Roy, Arpita
    Park, Youngja
    Pan, Shimei
    2019 IEEE 31ST INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2019), 2019, : 1568 - 1573
  • [7] A Knowledge-Enriched Ensemble Method for Word Embedding and Multi-Sense Embedding
    Fang, Lanting
    Luo, Yong
    Feng, Kaiyu
    Zhao, Kaiqi
    Hu, Aiqun
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (06) : 5534 - 5549
  • [8] Improving Twitter Sentiment Classification Using Topic-Enriched Multi-Prototype Word Embeddings
    Ren, Yafeng
    Zhang, Yue
    Zhang, Meishan
    Ji, Donghong
    THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 3038 - 3044
  • [9] Sense-Aware Semantic Analysis: A Multi-Prototype Word Representation Model Using Wikipedia
    Wu, Zhaohui
    Giles, C. Lee
    PROCEEDINGS OF THE TWENTY-NINTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2015, : 2188 - 2194
  • [10] Quantum word embedding for machine learning
    Nguyen, Phuong-Nam
    PHYSICA SCRIPTA, 2024, 99 (08)