Evaluating Unsupervised Text Embeddings on Software User Feedback

被引:11
|
作者
Devine, Peter [1 ]
Koh, Yun Sing [1 ]
Blincoe, Kelly [1 ]
机构
[1] Univ Auckland, Auckland, New Zealand
关键词
REVIEWS; MODELS;
D O I
10.1109/REW53955.2021.00020
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
User feedback on software products has been shown to be useful for development and can be exceedingly abundant online. Many approaches have been developed to elicit requirements in different ways from this large volume of feedback, including the use of unsupervised clustering, underpinned by text embeddings. Methods for embedding text can vary significantly within the literature, highlighting the lack of a consensus as to which approaches are best able to cluster user feedback into requirements relevant groups. This work proposes a methodology for comparing text embeddings of user feedback using existing labelled datasets. Using 7 diverse datasets from the literature, we apply this methodology to evaluate both established text embedding techniques from the user feedback analysis literature (including topic modelling and word embeddings) as well as text embeddings from state of the art deep text embedding models. Results demonstrate that text embeddings produced by state of the art models, most notably the Universal Sentence Encoder (USE), group feedback with similar requirements relevant characteristics together better than other evaluated techniques across all seven datasets. These results can help researchers select appropriate embedding techniques when developing future unsupervised clustering approaches within user feedback analysis.
引用
收藏
页码:87 / 95
页数:9
相关论文
共 50 条
  • [31] Analysis of user-feedback as a tool for improving software quality
    Abookire, SA
    Martin, MT
    Teich, JM
    Kuperman, GJ
    Bates, DW
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2000, : 940 - 940
  • [33] CTRLEval: An Unsupervised Reference-Free Metric for Evaluating Controlled Text Generation
    Ke, Pei
    Zhou, Hao
    Lin, Yankai
    Li, Peng
    Zhou, Jie
    Zhu, Xiaoyan
    Huang, Minlie
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 2306 - 2319
  • [34] Text Embeddings Reveal (Almost) As Much As Text
    Morris, John X.
    Kuleshov, Volodymyr
    Shmatikov, Vitaly
    Rush, Alexander M.
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 12448 - 12460
  • [35] Evaluating Software Quality Through User Reviews: The ISOftSentiment Tool
    Hou, Fang
    Feng, Liang
    Farshidi, Siamak
    Jansen, Slinger
    PRODUCT-FOCUSED SOFTWARE PROCESS IMPROVEMENT, PROFES 2024, 2025, 15452 : 75 - 91
  • [36] Evaluating Software Quality in Use using User Reviews Mining
    Leopairote, Warit
    Surarerks, Athasit
    Prompoon, Nakornthip
    2013 10TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER SCIENCE AND SOFTWARE ENGINEERING (JCSSE), 2013, : 257 - 262
  • [37] Evaluating User Experience of a Software Engineering Education Virtual Environment
    Fernandes, Filipe
    Castro, Diego
    Rodrigues, Claudia
    Werner, Claudia
    24TH SYMPOSIUM ON VIRTUAL AND AUGMENTED REALITY, SVR 2022, 2022, : 137 - 141
  • [38] New criterion for evaluating the design of software agents: user autonomy
    Tao, Qingping
    Cai, Qingsheng
    Xiaoxing Weixing Jisuanji Xitong/Mini-Micro Systems, 2000, 21 (02): : 148 - 149
  • [39] User preferences in evaluating usability of software product: A multicriteria approach
    Sikorski, M
    BIS'99: 3RD INTERNATIONAL CONFERENCE ON BUSINESS INFORMATION SYSTEMS, 1999, : 182 - 187
  • [40] Unsupervised Methods for the Study of Transformer Embeddings
    Saada, Mira Ait
    Role, Francois
    Nadif, Mohamed
    ADVANCES IN INTELLIGENT DATA ANALYSIS XIX, IDA 2021, 2021, 12695 : 287 - 300