Key points are not available for this paper at this time.
In-context learning (ICL), which teaches a large language model (LLM) to perform a task with few-shot demonstrations rather than adjusting the model parameters, has emerged as a strong paradigm for using LLMs. While early studies primarily used a fixed or random set of demonstrations for all test queries, recent research suggests that retrieving semantically similar demonstrations to the input from a pool of available demonstrations results in better performance. This work expands the applicability of retrieval-based ICL approaches along several dimensions. We extend the success of retrieval-based ICL to instruction-finetuned LLMs as well as Chain-of-Thought (CoT) prompting. While the prior work utilizes general Large Language Models (LLMs), such as GPT-3, we find that retrieved demonstrations also enhance instruction-finetuned LLMs. This insight implies that training data, despite being exposed during the fine-tuning phase, can still be effectively used through retrieval and in-context demonstrations during testing, resulting in superior outcomes when compared to utilizing no demonstrations or selecting them at random. For CoT, when the demonstrations contain reasoning chains, we get improvements by retrieving based on such chains. Finally, we train a task-specific demonstration retriever that outperforms off-the-shelf retrievers.
Building similarity graph...
Analyzing shared references across papers
Loading...
Man Luo
Xin Xu
Zhuyun Dai
Data Intelligence
Arizona State University
Google (United States)
Building similarity graph...
Analyzing shared references across papers
Loading...
Luo et al. (Mon,) studied this question.
www.synapsesocial.com/papers/68e61deab6db6435875afd09 — DOI: https://doi.org/10.3724/2096-7004.di.2024.0012
Synapse has enriched 5 closely related papers on similar clinical questions. Consider them for comparative context: