Efficient and Effective In-context Demonstration Selection with Coreset

arXiv — cs.CVThursday, November 13, 2025 at 5:00:00 AM
The recent publication of 'Efficient and Effective In-context Demonstration Selection with Coreset' on arXiv presents a significant advancement in the field of in-context learning (ICL) for large visual language models (LVLMs). The paper highlights the NP-hard nature of demonstration selection, which is critical for the effectiveness of ICL. Traditional methods often lead to inefficiencies, prompting the authors to propose a novel framework known as Coreset-based Dual Retrieval (CoDR). This innovative approach employs a cluster-pruning method to create a diverse coreset that aligns better with queries while maintaining diversity. Additionally, the dual retrieval mechanism enhances the selection process, achieving global demonstration selection efficiently. Experimental results indicate that CoDR significantly outperforms existing strategies, marking a notable improvement in ICL performance. This development not only addresses a key challenge in AI but also sets the stage for more effec…
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended Readings
Identifying and Analyzing Performance-Critical Tokens in Large Language Models
NeutralArtificial Intelligence
The paper titled 'Identifying and Analyzing Performance-Critical Tokens in Large Language Models' explores how large language models (LLMs) utilize in-context learning (ICL) for few-shot learning. It categorizes tokens in ICL prompts into content, stopword, and template tokens, aiming to identify those that significantly impact LLM performance. The study reveals that template and stopword tokens have a greater influence on performance than informative content tokens, challenging existing assumptions about human attention to informative words.
Learn to Select: Exploring Label Distribution Divergence for In-Context Demonstration Selection in Text Classification
PositiveArtificial Intelligence
The article discusses a novel approach to in-context learning (ICL) for text classification, emphasizing the importance of selecting appropriate demonstrations. Traditional methods often prioritize semantic similarity, neglecting label distribution alignment, which can impact performance. The proposed method, TopK + Label Distribution Divergence (L2D), utilizes a fine-tuned BERT-like small language model to generate label distributions and assess their divergence. This dual focus aims to enhance the effectiveness of demonstration selection in large language models (LLMs).