MPMA: Preference Manipulation Attack Against Model Context Protocol

arXiv — cs.CLWednesday, November 12, 2025 at 5:00:00 AM
The introduction of the MCP Preference Manipulation Attack (MPMA) highlights significant vulnerabilities in the Model Context Protocol (MCP), which standardizes how large language models (LLMs) access external data and tools. As the MCP gains traction, third-party customized versions pose security risks, enabling attackers to manipulate LLMs to favor their servers. This manipulation can yield economic benefits, such as revenue from paid services or advertising income from free servers. The attack's effectiveness is enhanced through methods like the Direct Preference Manipulation Attack (DPMA), which modifies tool names and descriptions to influence LLMs. However, DPMA's overt nature necessitated the development of the Genetic-based Advertising Preference Manipulation Attack (GAPMA), which balances effectiveness with stealth. The emergence of MPMA underscores the need for robust security measures in the rapidly evolving landscape of LLMs and their associated tools.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended Readings
Silenced Biases: The Dark Side LLMs Learned to Refuse
NegativeArtificial Intelligence
Safety-aligned large language models (LLMs) are increasingly used in sensitive applications where fairness is crucial. Evaluating their fairness is complex, often relying on standard question-answer methods that misinterpret refusal responses as indicators of fairness. This paper introduces the concept of silenced biases, which are unfair preferences hidden within the models' latent space, masked by safety-alignment. Previous methods have limitations, prompting the need for new approaches to uncover these biases effectively.
Fair In-Context Learning via Latent Concept Variables
PositiveArtificial Intelligence
The paper titled 'Fair In-Context Learning via Latent Concept Variables' explores the in-context learning (ICL) capabilities of large language models (LLMs) in handling tabular data. It highlights the potential for LLMs to inherit biases from pre-training data, which can lead to discrimination in high-stakes applications. The authors propose an optimal demonstration selection method using latent concept variables to enhance task adaptation and fairness, alongside data augmentation strategies to minimize correlations between sensitive variables and predictive outcomes.
Scaling Latent Reasoning via Looped Language Models
PositiveArtificial Intelligence
The article presents Ouro, a family of pre-trained Looped Language Models (LoopLM) designed to enhance reasoning capabilities during the pre-training phase. Unlike traditional models that rely on explicit text generation, Ouro incorporates iterative computation in latent space and an entropy-regularized objective for depth allocation. The models, Ouro 1.4B and 2.6B, demonstrate superior performance, matching results of larger state-of-the-art models while emphasizing improved knowledge manipulation rather than increased capacity.
Preference Orchestrator: Prompt-Aware Multi-Objective Alignment for Large Language Models
PositiveArtificial Intelligence
The article introduces the PReference Orchestrator (PRO), a framework designed to enhance the alignment of Large Language Models (LLMs) with diverse human preferences across multiple objectives. Traditional methods rely on manually set preference weights, which can hinder training efficiency and complicate user experience. PRO addresses these challenges by utilizing a lightweight preference adapter that automatically infers prompt-specific preference weights during both training and deployment, thereby improving performance and efficiency.
Modeling and Predicting Multi-Turn Answer Instability in Large Language Models
NeutralArtificial Intelligence
The paper titled 'Modeling and Predicting Multi-Turn Answer Instability in Large Language Models' discusses the evaluation of large language models (LLMs) in terms of their robustness during user interactions. The study employs multi-turn follow-up prompts to assess changes in model answers and accuracy dynamics using Markov chains. Results indicate vulnerabilities in LLMs, with a 10% accuracy drop for Gemini 1.5 Flash after a 'Think again' prompt over nine turns, and a 7.5% drop for Claude 3.5 Haiku with a reworded question. The findings suggest that accuracy can be modeled over time.
Pre-Attention Expert Prediction and Prefetching for Mixture-of-Experts Large Language Models
PositiveArtificial Intelligence
The paper titled 'Pre-Attention Expert Prediction and Prefetching for Mixture-of-Experts Large Language Models' introduces a method to enhance the efficiency of Mixture-of-Experts (MoE) Large Language Models (LLMs). The authors propose a pre-attention expert prediction technique that improves accuracy and reduces computational overhead by utilizing activations before the attention block. This approach aims to optimize expert prefetching, achieving about a 15% improvement in accuracy over existing methods.
From Fact to Judgment: Investigating the Impact of Task Framing on LLM Conviction in Dialogue Systems
NeutralArtificial Intelligence
The article investigates the impact of task framing on the conviction of large language models (LLMs) in dialogue systems. It explores how LLMs assess tasks requiring social judgment, contrasting their performance on factual queries with conversational judgment tasks. The study reveals that reframing a task can significantly alter an LLM's judgment, particularly under conversational pressure, highlighting the complexities of LLM decision-making in social contexts.
Multimodal Peer Review Simulation with Actionable To-Do Recommendations for Community-Aware Manuscript Revisions
PositiveArtificial Intelligence
A new interactive web-based system for multimodal peer review simulation has been introduced, aimed at enhancing manuscript revisions prior to submission. This system leverages large language models (LLMs) to integrate textual and visual information, improving the quality of reviews through retrieval-augmented generation (RAG) based on OpenReview data. It converts generated reviews into actionable to-do lists, providing structured guidance for authors and seamlessly integrating with existing academic writing platforms.