AgentDistill: Training-Free Agent Distillation with Generalizable MCP Boxes
Abstract
AgentDistill efficiently transfers knowledge from large language model-based agents to smaller ones using reusable, autonomously generated task-solving modules, enabling scalable and cost-effective intelligent agents.
While knowledge distillation has become a mature field for compressing large language models (LLMs) into smaller ones by aligning their outputs or internal representations, the distillation of LLM-based agents, which involve planning, memory, and tool use, remains relatively underexplored. Existing agent distillation methods typically replay full teacher trajectories or imitate step-by-step teacher tool usage, but they often struggle to train student agents to dynamically plan and act in novel environments. We propose AgentDistill, a novel, training-free agent distillation framework that enables efficient and scalable knowledge transfer via direct reuse of Model-Context-Protocols (MCPs), which are structured and reusable task-solving modules autonomously generated by teacher agents. The reuse of these distilled MCPs enables student agents to generalize their capabilities across domains and solve new problems with minimal supervision or human intervention. Experiments on biomedical and mathematical benchmarks demonstrate that our distilled student agents, built on small language models, can achieve performance comparable to advanced systems using large LLMs such as OctoTools (GPT-4o), highlighting the effectiveness of our framework in building scalable and cost-efficient intelligent agents.
Community
Good work. Could you provide the code?
@nuojohnchen I did a quick search on GitHub, and was wondering if is this the source code that you are looking for? It seems to match https://github.com/jtk16/AgentDistillation
Thanks for looking that up! I was looking for the MCP code part, which jtk16/AgentDistillation unfortunately doesn't have. It's all good though, I've figured it out. Thanks for the reply!
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper