"In-Context Learning" or: How I learned to stop worrying and love "Applied Information Retrieval"

被引:0
|
作者
Parry, Andrew [1 ]
Ganguly, Debasis [1 ]
Chandra, Manish [1 ]
机构
[1] Univ Glasgow, Glasgow, Lanark, Scotland
关键词
Large Language Models; In-Context Learning; Ranking Models; Query Performance Prediction;
D O I
10.1145/3626772.3657842
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the increasing ability of large language models (LLMs), incontext learning (ICL) has evolved as a new paradigm for natural language processing (NLP), where instead of fine-tuning the parameters of an LLM specific to a downstream task with labeled examples, a small number of such examples is appended to a prompt instruction for controlling the decoder's generation process. ICL, thus, is conceptually similar to a non-parametric approach, such as k-NN, where the prediction for each instance essentially depends on the local topology, i.e., on a localised set of similar instances and their labels (called few-shot examples). This suggests that a test instance in ICL is analogous to a query in IR, and similar examples in ICL retrieved from a training set relate to a set of documents retrieved from a collection in IR. While standard unsupervised ranking models can be used to retrieve these few-shot examples from a training set, the effectiveness of the examples can potentially be improved by re-defining the notion of relevance specific to its utility for the downstream task, i.e., considering an example to be relevant if including it in the prompt instruction leads to a correct prediction. With this task-specific notion of relevance, it is possible to train a supervised ranking model (e.g., a bi-encoder or cross-encoder), which potentially learns to optimally select the few-shot examples. We believe that the recent advances in neural rankers can potentially find a use case for this task of optimally choosing examples for more effective downstream ICL predictions.
引用
收藏
页码:14 / 25
页数:12
相关论文
共 50 条
  • [31] Eureka! or, How I learned to stop worrying and love the survey
    Kinginger, Celeste
    LANGUAGE TEACHING, 2021, 54 (01) : 38 - 46
  • [33] or: How I Learned to Stop Worrying and Love Cloud Computing
    Kashiwazaki, Hiroki
    IEEE 39TH ANNUAL COMPUTER SOFTWARE AND APPLICATIONS CONFERENCE WORKSHOPS (COMPSAC 2015), VOL 3, 2015, : 139 - 140
  • [34] The disappointed evangelist: or, how I learned to stop worrying and love ambiguity
    Adame, Alexandra
    REFLECTIVE PRACTICE, 2012, 13 (04) : 597 - 607
  • [35] Copenhagen computation: How I learned to stop worrying and love Bohr
    Mermin, ND
    IBM JOURNAL OF RESEARCH AND DEVELOPMENT, 2004, 48 (01) : 53 - 61
  • [36] How I Learned to Stop Worrying and Love Re-optimization
    Perron, Matthew
    Shang, Zeyuan
    Kraska, Tim
    Stonebraker, Michael
    2019 IEEE 35TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE 2019), 2019, : 1758 - 1761
  • [37] In Praise of Depth: or, How I Learned to Stop Worrying and Love the Hidden
    Landy, Joshua
    NEW LITERARY HISTORY, 2020, 51 (01) : 145 - 176
  • [38] Danse Macabre or: How I learned to stop worrying and love death
    Lolli Ghetti, Carlo Maria
    FORUM ITALICUM, 2024,
  • [39] Categories: How I Learned to Stop Worrying and Love Two Sorts
    Conradie, Willem
    Frittella, Sabine
    Palmigiano, Alessandra
    Piazzai, Michele
    Tzimoulis, Apostolos
    Wijnberg, Nachoem M.
    LOGIC, LANGUAGE, INFORMATION, AND COMPUTATION, 2016, 9803 : 145 - 164
  • [40] Technology and "the International' or: How I Learned to Stop Worrying and Love Determinism
    McCarthy, Daniel R.
    MILLENNIUM-JOURNAL OF INTERNATIONAL STUDIES, 2013, 41 (03): : 470 - 490