ai-agentsintermediate

Agent Eval Benchmark Tool

Compare coding agents head-to-head on custom tasks with pass rate, cost, time, and consistency metrics.

Install

/plugin install agent-eval-benchmark-tool@affaan-m

Requires Claude Code CLI.

Use cases

Teams evaluating which coding agent (Claude Code, Aider, etc.) to adopt by running reproducible task benchmarks.

Reviews

No reviews yet. Be the first to review this skill.

Stats

Installs0
GitHub Stars112.0k
Forks14594
LicenseMIT
UpdatedMar 27, 2026