Evaluation

What Is AgentBench?

A benchmark for evaluating LLM agents across interactive, multi-step environments that require planning, tool use, observation handling, and task completion.

What Is AgentBench?

AgentBench is a public agent-evaluation benchmark that tests whether large language models can act through multi-step environments, not just answer static prompts. It belongs to the evaluation family and appears in offline agent eval pipelines, benchmark comparisons, and production regression suites that replay traces. AgentBench-style tasks score planning, tool use, observation handling, and final success across environments such as web browsing, databases, operating systems, and games. In FutureAGI, teams pair that benchmark view with TrajectoryScore on real agent traces.

Why AgentBench Matters in Production LLM and Agent Systems

AgentBench matters because agents often fail before the final answer is written. A run can look plausible to a user while the agent chose the wrong tool, ignored an observation, repeated a step, or solved a database task by luck. If you ignore this benchmark class, the production failure mode is hidden trajectory failure: the answer may pass a shallow judge, but the path is unsafe, expensive, or impossible to reproduce.

Developers feel the pain first. They compare two model releases and see the same final-answer score, yet one release adds extra browser actions and retries. SREs see p99 latency rise, token-cost-per-trace increase, and tool timeout alerts cluster around a few task types. Product teams see users abandon long agent sessions because the agent wanders through irrelevant actions. Compliance reviewers get audit logs that show the agent touched a sensitive tool even though the final text sounded acceptable.

AgentBench is especially relevant for 2026 agent stacks because workflows now include planners, tool routers, retrieval, browser control, code execution, and handoffs. Unlike MT-Bench, which mainly compares conversational answers, AgentBench-style evaluation checks whether the system can keep state across an action loop. The useful production question is not “did the final answer sound good?” It is “did the agent take the right steps, in the right order, under the right constraints?”

How FutureAGI Maps AgentBench to TrajectoryScore

FutureAGI handles AgentBench as a benchmark-to-trace workflow anchored on the TrajectoryScore evaluator in fi.evals. The benchmark task becomes a dataset row: user goal, environment metadata, available tools, expected terminal condition, and optional reference actions. The agent run becomes a trace, with each action recorded as agent.trajectory.step through a traceAI integration such as traceAI-langchain or traceAI-openai-agents.

A real workflow: an engineering team ports AgentBench-style browser and database tasks into a 300-row golden dataset for a customer-support agent. Each row defines the task, allowed tools, expected resolution, and disallowed actions. FutureAGI runs the current prompt and a candidate prompt, stores the trace, then scores every run with TrajectoryScore. Engineers add ToolSelectionAccuracy when the failure depends on choosing the right API, and TaskCompletion when they need a clean goal-level pass/fail signal.

FutureAGI’s approach is to treat AgentBench as coverage evidence, not a release decision by itself. We’ve found that public agent benchmarks are most useful when their failures become product-specific regression rows. If TrajectoryScore drops on browser-navigation tasks while TaskCompletion stays flat, the engineer inspects the failing agent.trajectory.step, tightens the browser policy, adds a regression threshold, and reruns the eval before routing more traffic to the candidate model.

How to Measure or Detect AgentBench Performance

Measure AgentBench-style performance with task success, trace quality, and production side effects:

  • TrajectoryScore — the FutureAGI evaluator for comprehensive trajectory evaluation; use it as the headline score for multi-step agent runs.
  • ToolSelectionAccuracy — checks whether the agent selected the expected tool when the benchmark task has a known action path.
  • agent.trajectory.step — the trace field that locates which observation, action, or tool call caused the benchmark failure.
  • Dashboard signals — track eval-fail-rate-by-cohort, p99 latency, token-cost-per-trace, retry count, and tool-timeout rate for benchmark-like tasks.
  • User-feedback proxy — compare benchmark failures against thumbs-down rate, session abandonment, escalation rate, and manual QA notes.

Minimal Python:

from fi.evals import TrajectoryScore

metric = TrajectoryScore()
result = metric.evaluate(
    trajectory=run["trajectory"],
    task=run["task"],
    available_tools=run["tools"],
)
print(result.score, result.reason)

Common Mistakes

  • Treating AgentBench as a production acceptance test. Its environments are useful stressors, but they are not your tools, policies, latency limits, or customers.
  • Scoring only final success. Agent runs need trace-level checks for wrong intermediate actions, repeated steps, ignored observations, and unsafe tool access.
  • Comparing models without fixed prompts and tool schemas. AgentBench-style results drift when the action space changes between runs.
  • Ignoring cost and latency. A model can improve task success while doubling token-cost-per-trace or p99 latency.
  • Using public benchmark rank as a route policy. Route production traffic from task-specific evals, not from a generic agent table.

Frequently Asked Questions

What is AgentBench?

AgentBench is a benchmark for evaluating whether LLMs can act through multi-step environments such as web browsing, databases, operating systems, knowledge graphs, and games. FutureAGI maps AgentBench-style tasks to trajectory scores, traces, and regression evals for production agents.

How is AgentBench different from trajectory score?

AgentBench is a benchmark suite: it defines task environments and comparison conditions. Trajectory score is a FutureAGI evaluator metric that scores a captured agent run, so it is better suited for production regression and trace analysis.

How do you measure AgentBench?

Run the benchmark tasks, capture each agent trajectory, then score the run with FutureAGI's TrajectoryScore and related signals such as ToolSelectionAccuracy. The trace field agent.trajectory.step helps identify which action caused a failure.