BiLSTM-CRF Manipuri NER with Character-Level Word Representation

被引:6
|
作者
Jimmy, Laishram [1 ]
Nongmeikappam, Kishorjit [2 ]
Naskar, Sudip Kumar [3 ]
机构
[1] Manipur Tech Univ, Imphal, Manipur, India
[2] Indian Inst Informat Technol Manipur, Imphal, Manipur, India
[3] Jadavpur Univ, Kolkata, W Bengal, India
关键词
Manipuri; Named entity recognition and classification; LSTM; CRF; Embeddings; Deep neural networks; Recurrent neural networks; NAMED ENTITY RECOGNITION; MODEL;
D O I
10.1007/s13369-022-06933-z
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Named Entity Recognition and Classification (NER) serves as a foundation for many natural language processing tasks such as question answering, text summarization, news/document clustering and machine translation. Manipuri's early NER systems are based on machine learning approaches and employ handcrafted morphological features and domain-specific rules. The domain-specific rules for Manipuri NER are hard to extract as the language is highly agglutinative, inflectional and falls in the category of low resource language. In recent years, deep learning, empowered by continuous vector representation and semantic composition through non-linear processing, has been employed in the various NER task yielding state-of-the accuracy. In this paper, we propose a Manipuri NER model using Bidirectional Long Short Term Memory (BiLSTM) deep neural network in unison with an embedding technique. The embedding technique is a BiLSTM character-level word representation in conjunction with word embedding, which acts as a feature for the Bi-LSTM NER model. The proposed model also employs a Conditional Random Field (CRF) classifier to capture the dependency among output NER tags. Various Gradient Descent (GD) optimizers for the neural model were experimented with to establish an efficient GD optimizer for accurate NER. The NER model with RMSprop GD optimizer achieved an F-Score measure of approximately 98.19% at learning rate eta = 0.001 and with decay constant of rho = 0.9. Further, while performing an intrinsic evaluation on the word embedding, it is found that the proposed embedding technique as a feature can capture the semantic and syntactic rule of the language with 88.14% average clustering accuracy for all NE classes.
引用
收藏
页码:1715 / 1734
页数:20
相关论文
共 38 条
  • [21] Extracting Drug-Drug Interactions with Word and Character-Level Recurrent Neural Networks
    Kavuluru, Ramakanth
    Rios, Anthony
    Tung Tran
    2017 IEEE INTERNATIONAL CONFERENCE ON HEALTHCARE INFORMATICS (ICHI), 2017, : 5 - 12
  • [22] Character-Level Dependency Model for Joint Word Segmentation, POS Tagging, and Dependency Parsing in Chinese
    Guo, Zhen
    Zhang, Yujie
    Su, Chen
    Xu, Jinan
    Isahara, Hitoshi
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2016, E99D (01): : 257 - 264
  • [23] Character-Level Chinese Dependency Parsing via Modeling Latent Intra-Word Structure
    Hou, Yang
    Li, Zhenghua
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 2943 - 2956
  • [24] Joint Word- and Character-level Embedding CNN-RNN Models for Punctuation Restoration
    Tundik, Mate Akos
    Szaszak, Gyorgy
    2018 9TH IEEE INTERNATIONAL CONFERENCE ON COGNITIVE INFOCOMMUNICATIONS (COGINFOCOM), 2018, : 135 - 140
  • [25] Using Character-Level Sequence-to-Sequence Model for Word Level Text Generation to Enhance Arabic Speech Recognition
    Azim, Mona A.
    Hussein, Wedad
    Badr, Nagwa L.
    IEEE ACCESS, 2023, 11 : 91173 - 91183
  • [26] Thai Named Entity Recognition Using Bi-LSTM-CRF with Word and Character Representation
    Thattinaphanich, Suphanut
    Prom-on, Santitham
    PROCEEDINGS OF THE 2019 4TH INTERNATIONAL CONFERENCE ON INFORMATION TECHNOLOGY (INCIT): ENCOMPASSING INTELLIGENT TECHNOLOGY AND INNOVATION TOWARDS THE NEW ERA OF HUMAN LIFE, 2019, : 149 - 154
  • [27] Reading strategy of Hong Kong school-aged children: The development of word-level and character-level processing
    Chu, MMK
    Leung, MT
    APPLIED PSYCHOLINGUISTICS, 2005, 26 (04) : 505 - 520
  • [28] Character-level word encoding deep learning model for combating cyber threats in phishing URL detection
    Alshehri, Mohammed
    Abugabah, Ahed
    Algarni, Abdullah
    Almotairi, Sultan
    COMPUTERS & ELECTRICAL ENGINEERING, 2022, 100
  • [29] LSTM Easy-first Dependency Parsing with Pre-trained Word Embeddings and Character-level Word Embeddings in Vietnamese
    Binh Duc Nguyen
    Kiet Van Nguyen
    Ngan Luu-Thuy Nguyen
    PROCEEDINGS OF 2018 10TH INTERNATIONAL CONFERENCE ON KNOWLEDGE AND SYSTEMS ENGINEERING (KSE), 2018, : 187 - 192
  • [30] Machine learning for financial transaction classification across companies using character-level word embeddings of text fields
    Jorgensen, Rasmus Kaer
    Igel, Christian
    INTELLIGENT SYSTEMS IN ACCOUNTING FINANCE & MANAGEMENT, 2021, 28 (03): : 159 - 172