Agent Benchmarks
Viewer • Updated • 692 • 424 • 15Note SciBench is a novel benchmark for college-level scientific problems sourced from instructional textbooks
google/frames-benchmark
Viewer • Updated • 824 • 2.28k • 212Note 824 challenging multi-hop questions requiring information from 2-15 Wikipedia articles. If we adapt this to a search agent with no RAG component it's super interesting.
gaia-benchmark/GAIA
Updated • 9.47k • 376
HuggingFaceH4/MATH-500
Viewer • Updated • 500 • 60.2k • 158Note can be nicely adapted to an agentic setting with code
smolagents/browse_comp
Viewer • Updated • 1.27k • 193 • 1Note A simple and challenging benchmark that measures the ability of AI agents to locate hard-to-find information online
THUDM/ComplexFuncBench
Updated • 134 • 7Note The ComplexFuncBench dataset encompass 1,000 complex function calling samples from five aspects: (1) Function calling with multi-step in single turn; (2) Function calling with user-provided constraints; (3) Function calling that requires parameter value reasoning from implicit information; (4) Function calling with long parameter values that exceed 500 tokens; and (5) Function calling with 128k long-context length.
galileo-ai/agent-leaderboard
Viewer • Updated • 1.28k • 346 • 29Note Comprehensive evaluation across multiple domains and interaction types by leveraging diverse datasets: BFCL: Mathematics, Entertainment, Education, and Academic Domains τ-bench: Retail and Airline Industry Scenarios xLAM: Cross-domain Data Generation (21 Domains) ToolACE: API Interactions across 390 Domains
THUDM/SWE-Dev-train
Viewer • Updated • 20.1k • 165 • 6Note Training traces that helped a 32B model achieve gpt-4o level performance on SWE-bench. OpenHands framework
SWE-Gym/SWE-Gym
Viewer • Updated • 2.44k • 31.6k • 14