perf-benchmarker
Use when running performance benchmarks, establishing baselines, or validating regressions with sequential runs. Enforces 60s minimum runs (30s only for binary search) and no parallel benchmarks.
Packaged view
This page reorganizes the original catalog entry around fit, installability, and workflow context first. The original raw source lives below.
Install command
npx @skill-hub/cli install agent-sh-agentsys-perf-benchmarker
Repository
Skill path: .kiro/skills/perf-benchmarker
Use when running performance benchmarks, establishing baselines, or validating regressions with sequential runs. Enforces 60s minimum runs (30s only for binary search) and no parallel benchmarks.
Open repositoryBest for
Primary workflow: Ship Full Stack.
Technical facets: Full Stack.
Target audience: everyone.
License: Unknown.
Original source
Catalog source: SkillHub Club.
Repository owner: agent-sh.
This is still a mirrored public skill entry. Review the repository before installing into production workflows.
What it helps with
- Install perf-benchmarker into Claude Code, Codex CLI, Gemini CLI, or OpenCode workflows
- Review https://github.com/agent-sh/agentsys before adding perf-benchmarker to shared team environments
- Use perf-benchmarker for development workflows
Works across
Favorites: 0.
Sub-skills: 0.
Aggregator: No.
Original source / Raw SKILL.md
---
name: perf-benchmarker
description: "Use when running performance benchmarks, establishing baselines, or validating regressions with sequential runs. Enforces 60s minimum runs (30s only for binary search) and no parallel benchmarks."
version: 5.1.0
argument-hint: "[command] [duration]"
---
# perf-benchmarker
Run sequential benchmarks with strict duration rules.
Follow `docs/perf-requirements.md` as the canonical contract.
## Parse Arguments
```javascript
const args = '$ARGUMENTS'.split(' ').filter(Boolean);
const command = args.find(a => !a.match(/^\d+$/)) || '';
const duration = parseInt(args.find(a => a.match(/^\d+$/)) || '60', 10);
```
## Required Rules
- Benchmarks MUST run sequentially (never parallel).
- Minimum duration: 60s per run (30s only for binary search).
- Warmup: 10s minimum before measurement.
- Re-run anomalies.
## Output Format
```
command: <benchmark command>
duration: <seconds>
warmup: <seconds>
results: <metrics summary>
notes: <anomalies or reruns>
```
## Output Contract
Benchmarks MUST emit a JSON metrics block between markers:
```
PERF_METRICS_START
{"scenarios":{"low":{"latency_ms":120},"high":{"latency_ms":450}}}
PERF_METRICS_END
```
## Constraints
- No short runs unless binary-search phase.
- Do not change code while benchmarking.