58
17

Cut the Crap: An Economical Communication Pipeline for LLM-based Multi-Agent Systems

Abstract

Recent advancements in large language model (LLM)-powered agents have shown that collective intelligence can significantly outperform individual capabilities, largely attributed to the meticulously designed inter-agent communication topologies. Though impressive in performance, existing multi-agent pipelines inherently introduce substantial token overhead, as well as increased economic costs, which pose challenges for their large-scale deployments. In response to this challenge, we propose an economical, simple, and robust multi-agent communication framework, termed AgentPrune\texttt{AgentPrune}, which can seamlessly integrate into mainstream multi-agent systems and prunes redundant or even malicious communication messages. Technically, AgentPrune\texttt{AgentPrune} is the first to identify and formally define the \textit{communication redundancy} issue present in current LLM-based multi-agent pipelines, and efficiently performs one-shot pruning on the spatial-temporal message-passing graph, yielding a token-economic and high-performing communication topology. Extensive experiments across six benchmarks demonstrate that AgentPrune\texttt{AgentPrune} \textbf{(I)} achieves comparable results as state-of-the-art topologies at merely \5.6costcomparedtotheir cost compared to their \43.743.7, \textbf{(II)} integrates seamlessly into existing multi-agent frameworks with 28.1%72.8%28.1\%\sim72.8\%\downarrow token reduction, and \textbf{(III)} successfully defend against two types of agent-based adversarial attacks with 3.5%10.8%3.5\%\sim10.8\%\uparrow performance boost.

View on arXiv
Comments on this paper

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.