MURMUR: Using cross-user chatter to break collaborative language agents in groups

arXiv — cs.CLTuesday, November 25, 2025 at 5:00:00 AM
  • A recent study introduces MURMUR, a framework that reveals vulnerabilities in collaborative language agents through cross-user poisoning (CUP) attacks. These attacks exploit the lack of isolation in user interactions within multi-user environments, allowing adversaries to manipulate shared states and trigger unintended actions by the agents. The research validates these attacks on popular multi-user systems, highlighting a significant security concern in the evolving landscape of AI collaboration.
  • The implications of this research are profound, as it underscores the need for enhanced security measures in language models that operate in group settings. As these models become integral to collaborative tasks, understanding and mitigating risks like CUP is crucial for maintaining user trust and ensuring the reliability of AI systems in shared workspaces.
  • This development reflects a broader trend in AI research, where the balance between collaboration and security is increasingly scrutinized. While frameworks like Multi-Agent Collaborative Filtering aim to improve user interactions and recommendations, the potential for exploitation through vulnerabilities like CUP raises critical questions about privacy and the integrity of AI systems, necessitating ongoing dialogue and innovation in safeguarding user data.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended apps based on your readingExplore all apps
Continue Readings
FanarGuard: A Culturally-Aware Moderation Filter for Arabic Language Models
PositiveArtificial Intelligence
A new moderation filter named FanarGuard has been introduced, designed specifically for Arabic language models. This bilingual filter assesses both safety and cultural alignment in Arabic and English, utilizing a dataset of over 468,000 prompt-response pairs evaluated by human raters. The development aims to address the shortcomings of existing moderation systems that often neglect cultural nuances.
Word-level Annotation of GDPR Transparency Compliance in Privacy Policies using Large Language Models
PositiveArtificial Intelligence
A new study presents a modular large language model (LLM)-based pipeline for word-level annotation of privacy policies, focusing on compliance with GDPR transparency requirements. This approach aims to address the challenges of manual audits and the limitations of existing automated methods by providing fine-grained, context-aware annotations across 21 transparency requirements.
Beyond Multiple Choice: Verifiable OpenQA for Robust Vision-Language RFT
PositiveArtificial Intelligence
A new framework called ReVeL (Rewrite and Verify by LLM) has been proposed to enhance the multiple-choice question answering (MCQA) format used in evaluating multimodal language models. This framework transforms MCQA into open-form questions while ensuring answers remain verifiable, addressing issues of answer guessing and unreliable accuracy metrics during reinforcement fine-tuning (RFT).
Multi-Agent Collaborative Filtering: Orchestrating Users and Items for Agentic Recommendations
PositiveArtificial Intelligence
The Multi-Agent Collaborative Filtering (MACF) framework has been proposed to enhance agentic recommendations by utilizing large language model (LLM) agents that can interact with users and suggest relevant items based on collaborative signals from user-item interactions. This approach aims to improve the effectiveness of recommendation systems beyond traditional single-agent workflows.
From Competition to Coordination: Market Making as a Scalable Framework for Safe and Aligned Multi-Agent LLM Systems
PositiveArtificial Intelligence
A new market-making framework for coordinating multi-agent large language model (LLM) systems has been introduced, addressing challenges in trustworthiness and accountability as these models interact as agents. This framework enables agents to trade probabilistic beliefs, aligning local incentives with collective goals to achieve truthful outcomes without external enforcement.
Understanding and Mitigating Over-refusal for Large Language Models via Safety Representation
NeutralArtificial Intelligence
A recent study has highlighted the issue of over-refusal in large language models (LLMs), which occurs when these models excessively decline to generate outputs due to safety concerns. The research proposes a new approach called MOSR, which aims to balance safety and usability by addressing the representation of safety in LLMs.
For Those Who May Find Themselves on the Red Team
NeutralArtificial Intelligence
A recent position paper emphasizes the need for literary scholars to engage with research on large language model (LLM) interpretability, suggesting that the red team could serve as a platform for this ideological struggle. The paper argues that current interpretability standards are insufficient for evaluating LLMs.
ARQUSUMM: Argument-aware Quantitative Summarization of Online Conversations
PositiveArtificial Intelligence
A new framework called ARQUSUMM has been introduced to enhance the summarization of online conversations by focusing on the argumentative structure within discussions, particularly on platforms like Reddit. This approach aims to quantify argument strength and clarify the claim-reason relationships in conversations.