MCP-Universe: Benchmarking Large Language Models with
Real-World Model Context Protocol Servers
The first comprehensive benchmark specifically designed to evaluate LLMs in realistic and challenging tasks through interaction with real-world MCP servers across 6 core domains and 231 tasks.
Abstract
The Model Context Protocol (MCP) has emerged as a transformative standard for connecting large language models (LLMs) to external data sources and tools, rapidly gaining adoption across major AI providers and development platforms. However, existing benchmarks are overly simplistic and fail to capture real application challenges such as long-horizon reasoning and large, unfamiliar tool spaces.
To address this critical gap, we introduce MCP-Universe, the first comprehensive benchmark specifically designed to evaluate LLMs in realistic and hard tasks through interaction with real-world MCP servers. Our benchmark encompasses 6 core domains spanning 11 different MCP servers: Location Navigation, Repository Management, Financial Analysis, 3D Design, Browser Automation, and Web Searching. To ensure rigorous evaluation, we implement execution-based evaluators, including format evaluators for agent format compliance, static evaluators for time-invariant content matching, and dynamic evaluators that automatically retrieve real-time ground truth for temporally sensitive tasks.
Through extensive evaluation of leading LLMs, we find that even top-performing models such as GPT-5 (43.72% success rate), Grok-4 (33.33% success rate) and Claude-4.0-Sonnet (29.44% success rate) exhibit significant performance limitations. Beyond evaluation, we open-source our extensible evaluation framework with UI support, enabling researchers and practitioners to seamlessly integrate new agents and MCP servers while fostering innovation in the rapidly evolving MCP ecosystem.

Figure 1: Example from MCP-Universe illustrating realistic challenges, including real-world tool usage, long-horizon multi-turn tool calls, long context windows, scattered evidence, and large tool spaces. Unlike prior work, MCP-Universe is grounded in real-world MCP servers connected to actual data sources and environments.
Key Statistics
Benchmark Domains
Location Navigation
Real-world geospatial navigation tasks involving complex location queries, route planning, and geographic point calculations with actual map data.
Web Searching
Advanced web search tasks requiring multi-step information retrieval, synthesis, and real-time data processing from various sources.
Browser Automation
Complex browser automation tasks involving real-time web interactions, request submissions, and dynamic content extraction.
3D Design
Three-dimensional modeling and design tasks using real Blender software tools with geometric constraints and design specifications.
Financial Analysis
Real-time financial data analysis, quantitative investing, market research, and investment calculations involving temporal dynamics and live market data.
Repository Management
Version control workflows, code repository management, and collaborative development tasks across different platforms like GitHub.
Benchmark Results
Performance comparison across different LLMs and Agents on MCP-Universe benchmark. Success Rate (SR), Average Evaluator Score, and Average Steps reported.
Model | Location Navigation |
Repository Management |
Financial Analysis |
3D Designing |
Browser Automation |
Web Searching |
Average Evaluator Score |
Average Steps |
Overall Success Rate |
---|---|---|---|---|---|---|---|---|---|
GPT-5 | 33.33 | 30.30 | 67.50 | 52.63 | 35.90 | 45.45 | 60.23 | 8.22 | 43.72 |
Grok-4 | 28.89 | 12.12 | 40.00 | 26.32 | 41.03 | 41.82 | 49.01 | 7.75 | 33.33 |
Claude-4.0-Sonnet | 22.22 | 12.12 | 55.00 | 26.32 | 38.46 | 21.82 | 50.61 | 7.46 | 29.44 |
o3 | 26.67 | 6.06 | 40.00 | 26.32 | 25.64 | 29.09 | 38.95 | 4.82 | 26.41 |
o4-mini | 26.67 | 18.18 | 40.00 | 36.84 | 23.08 | 18.18 | 40.38 | 7.90 | 25.97 |
Claude-3.7-Sonnet | 13.33 | 18.18 | 40.00 | 36.84 | 23.08 | 21.82 | 40.36 | 7.16 | 24.24 |
Gemini-2.5-Pro | 13.33 | 12.12 | 50.00 | 21.05 | 25.64 | 12.73 | 36.93 | 6.98 | 22.08 |
Gemini-2.5-Flash | 15.56 | 12.12 | 37.50 | 21.05 | 30.77 | 14.55 | 33.99 | 8.26 | 21.65 |
GPT-4.1 | 8.89 | 6.06 | 40.00 | 26.32 | 23.08 | 10.91 | 41.32 | 5.24 | 18.18 |
GPT-4o | 8.89 | 9.09 | 35.00 | 26.32 | 12.82 | 9.09 | 37.03 | 6.03 | 15.58 |
GLM-4.5 | 17.78 | 9.09 | 50.00 | 26.32 | 15.38 | 27.27 | 41.16 | 7.33 | 24.68 |
Kimi-K2 | 11.11 | 9.09 | 47.50 | 15.79 | 15.38 | 14.55 | 35.10 | 6.07 | 19.05 |
Qwen3-Coder | 8.89 | 3.03 | 50.00 | 26.32 | 25.64 | 10.91 | 37.78 | 7.78 | 19.91 |
Qwen3-235B | 11.11 | 9.09 | 50.00 | 15.79 | 15.38 | 9.09 | 38.53 | 5.74 | 18.18 |
DeepSeek-V3 | 11.11 | 6.06 | 30.00 | 26.32 | 12.82 | 7.27 | 35.82 | 5.06 | 14.29 |
GPT-OSS-120B | 6.67 | 6.06 | 35.00 | 10.53 | 5.13 | 5.45 | 26.34 | - | 11.26 |
Key Findings
Long-Context Challenge
Token count increases rapidly with interaction steps, often leading to context overflow and degraded performance in multi-step tasks requiring extensive reasoning.
Unknown-Tools Challenge
LLM agents often lack familiarity with precise usage patterns, parameter specifications, and expected behaviors of diverse MCP servers.
Cross-Domain Variations
Models show markedly different success rates across application domains, suggesting domain-specific optimization needs and knowledge gaps.