Item Categorization Algorithm Based on Improved Text Representation

被引:0
|
作者
Zhenchao T. [1 ]
Jing M. [1 ]
机构
[1] College of Economics and Management, Nanjing University of Aeronautics and Astronautics, Nanjing
基金
中国国家自然科学基金;
关键词
Item Categorization; Metric Learning; Multitasking Learning; Text Classification; Text Representation;
D O I
10.11925/infotech.2096-3467.2021.0958
中图分类号
学科分类号
摘要
[Objective] This paper proposes a new model to improve the traditional text classifiers which tend to misclassify commodity titles with different labels and similar modifiers. [Methods] First, we designed the text discriminator as an auxiliary task, which took the normalized Euclidean distance of different label text vectors as the loss function. Then, we utilized the cross-entropy loss function of the traditional text classification to the new text encoder. Finally, we generated text representation with sufficient discrimination for different categories of commodity texts, and constructed the ITR-BiLSTM-Attention model. [Results] Compared with the BiLSTM Attention model without text discriminator, the proposed model’s accuracy, precision, recall and F1 values improved by 1.84%, 2.31%, 2.88% and 2.82%, respectively. Compared with the Cos-BiLSTM-Attention model, our new model improved accuracy, precision, recall and F1 values by 0.53%, 0.54%, 1.21% and 1.01%, respectively. [Limitations] The impacts of different sampling methods on the model were not tested. We did not conduct experiment on a larger data set. [Conclusions] The text discriminator auxiliary task designed in this paper can improve the text representation generated by the text encoder. The item categorization model based on improved text representation was more effective than the traditional ones. © 2022, Chinese Academy of Sciences. All rights reserved.
引用
收藏
页码:34 / 43
页数:9
相关论文
共 38 条
  • [1] He Bo, Ma Jing, Li Chi, Research on Commodity Text Classification Based on Fusion Features, Information Studies: Theory & Application, 43, 11, pp. 162-168, (2020)
  • [2] Li Xiaofeng, Ma Jing, Li Chi, Et al., Identifying Commodity Names Based on XGBoost Model, Data Analysis and Knowledge Discovery, 3, 7, pp. 34-41, (2019)
  • [3] Wan Jiashan, Wu Yunzhi, Review of Text Classification Research Based on Deep Learning, Journal of Tianjin University of Technology, 37, 2, pp. 41-47, (2021)
  • [4] Ohashi S, Takayama J, Kajiwara T, Et al., Text Classification with Negative Supervision, Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 351-357, (2020)
  • [5] Shen D H, Wang G Y, Wang W L, Et al., Baseline Needs More Love: On Simple Word-Embedding-Based Models and Associated Pooling Mechanisms, Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, pp. 440-450, (2018)
  • [6] Yang Z C, Yang D Y, Dyer C, Et al., Hierarchical Attention Networks for Document Classification, Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 1480-1489, (2016)
  • [7] Qin Q, Hu W P, Liu B., Feature Projection for Improved Text Classification, Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 8161-8171, (2020)
  • [8] Hochreiter S, Schmidhuber J., Long Short-Term Memory, Neural Computation, 9, 8, pp. 1735-1780, (1997)
  • [9] Bengio Y, Ducharme R, Vincent P, Et al., A Neural Probabilistic Language Model, Journal of Machine Learning Research, 3, pp. 1137-1155, (2003)
  • [10] Mikolov T, Chen K, Corrado G, Et al., Efficient Estimation of Word Representations in Vector Space