Concept generalization and fusion for abstractive sentence generation

被引:8
|
作者
Belkebir, Riadh [1 ]
Guessoum, Ahmed [1 ]
机构
[1] Univ Sci & Technol Houari Boumediene, Dept Comp Sci, Artificial Intelligence Res Lab, Nat Language Proc & Machine Learning Res Grp, Algiers 16111, Algeria
关键词
Artificial intelligence; Natural language processing; Text summarization; Concept fusion; Concept generalization; WordNet; COMPRESSION;
D O I
10.1016/j.eswa.2016.01.007
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Text summarization is either extractive or abstractive. Extractive summarization is to select the most salient pieces of information (words, phrases, and/or sentences) from a source document without adding any external information. Abstractive summarization allows an internal representation of the source document so as to produce a faithful summary of the source. In this case, external text can be inserted into the generated summary. Because of the complexity of the abstractive approach, the vast majority of work in text summarization has adopted an extractive approach. In this work, we focus on concepts fusion and generalization, i.e. where different concepts appearing in a sentence can be replaced by one concept which covers the meanings of all of them. This is one operation that can be used as part of an abstractive text summarization system. The main goal of this contribution is to enrich the research efforts on abstractive text summarization with a novel approach that allows the generalization of sentences using semantic resources. This work should be useful in intelligent systems more generally since it introduces a means to shorten sentences by producing more general (hence abstractions of the) sentences. It could be used, for instance, to display shorter texts in applications for mobile devices. It should also improve the quality of the generated text summaries by mentioning key (general) concepts. One can think of using the approach in reasoning systems where different concepts appearing in the same context are related to one another with the aim of finding a more general representation of the concepts. This could be in the context of Goal Formulation, expert systems, scenario recognition, and cognitive reasoning more generally. We present our methodology for the generalization and fusion of concepts that appear in sentences. This is achieved through (1) the detection and extraction of what we define as generalizable sentences and (2) the generation and reduction of the space of generalization versions. We introduce two approaches we have designed to select the best sentences from the space of generalization versions. Using four NLTK1 corpora, the first approach estimates the "acceptability" of a given generalization version. The second approach is Machine Learning-based and uses contextual and specific features. The recall, precision and Fl-score measures resulting from the evaluation of the concept generalization and fusion approach are presented. (C) 2016 Elsevier Ltd. All rights reserved.
引用
收藏
页码:43 / 56
页数:14
相关论文
共 50 条
  • [31] GEMINI: Controlling The Sentence-Level Summary Style in Abstractive Text Summarization
    Bao, Guangsheng
    Ou, Zebin
    Zhang, Yue
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 831 - 842
  • [32] Causal Intervention for Abstractive Related Work Generation
    Liu, Jiachang
    Zhang, Qi
    Shi, Chongyang
    Naseem, Usman
    Wang, Shoujin
    Hu, Liang
    Tsang, Ivor W.
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 2148 - 2159
  • [33] Abstractive text summarization based on deep learning and semantic content generalization
    Kouris, Panagiotis
    Alexandridis, Georgios
    Stafylopatis, Andreas
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 5082 - 5092
  • [34] Multi-level shared-weight encoding for abstractive sentence summarization
    Daisy Monika Lal
    Krishna Pratap Singh
    Uma Shanker Tiwary
    Neural Computing and Applications, 2022, 34 : 2965 - 2981
  • [35] Multi-level shared-weight encoding for abstractive sentence summarization
    Lal, Daisy Monika
    Singh, Krishna Pratap
    Tiwary, Uma Shanker
    Neural Computing and Applications, 2022, 34 (04) : 2965 - 2981
  • [36] Multi-level shared-weight encoding for abstractive sentence summarization
    Lal, Daisy Monika
    Singh, Krishna Pratap
    Tiwary, Uma Shanker
    NEURAL COMPUTING & APPLICATIONS, 2022, 34 (04): : 2965 - 2981
  • [37] FAR-ASS: Fact-aware reinforced abstractive sentence summarization
    Zhang, Mengli
    Zhou, Gang
    Yu, Wanting
    Liu, Wenfen
    INFORMATION PROCESSING & MANAGEMENT, 2021, 58 (03)
  • [38] Word Segmentation for Arabic Abstractive Headline Generation
    Abdelaziz, Yaser O.
    El-Beltagy, Samhaa R.
    2021 4TH INTERNATIONAL CONFERENCE ON INFORMATION AND COMPUTER TECHNOLOGIES (ICICT 2021), 2021, : 59 - 63
  • [39] Automatic Chinese Abstractive Summarization With Topical Keywords Fusion
    Hou L.-W.
    Hu P.
    Cao W.-L.
    Zidonghua Xuebao/Acta Automatica Sinica, 2019, 45 (03): : 530 - 539
  • [40] Treatment and generalization of complex sentence production in agrammatism
    Ballard, KJ
    Thompson, CK
    JOURNAL OF SPEECH LANGUAGE AND HEARING RESEARCH, 1999, 42 (03): : 690 - 707