Back to skills
SkillHub ClubShip Full StackFull Stack

perf-benchmarker

Use when running performance benchmarks, establishing baselines, or validating regressions with sequential runs. Enforces 60s minimum runs (30s only for binary search) and no parallel benchmarks.

Packaged view

This page reorganizes the original catalog entry around fit, installability, and workflow context first. The original raw source lives below.

Stars
610
Hot score
99
Updated
March 20, 2026
Overall rating
C3.7
Composite score
3.7
Best-practice grade
S96.0

Install command

npx @skill-hub/cli install agent-sh-agentsys-perf-benchmarker

Repository

agent-sh/agentsys

Skill path: .kiro/skills/perf-benchmarker

Use when running performance benchmarks, establishing baselines, or validating regressions with sequential runs. Enforces 60s minimum runs (30s only for binary search) and no parallel benchmarks.

Open repository

Best for

Primary workflow: Ship Full Stack.

Technical facets: Full Stack.

Target audience: everyone.

License: Unknown.

Original source

Catalog source: SkillHub Club.

Repository owner: agent-sh.

This is still a mirrored public skill entry. Review the repository before installing into production workflows.

What it helps with

  • Install perf-benchmarker into Claude Code, Codex CLI, Gemini CLI, or OpenCode workflows
  • Review https://github.com/agent-sh/agentsys before adding perf-benchmarker to shared team environments
  • Use perf-benchmarker for development workflows

Works across

Claude CodeCodex CLIGemini CLIOpenCode

Favorites: 0.

Sub-skills: 0.

Aggregator: No.

Original source / Raw SKILL.md

---
name: perf-benchmarker
description: "Use when running performance benchmarks, establishing baselines, or validating regressions with sequential runs. Enforces 60s minimum runs (30s only for binary search) and no parallel benchmarks."
version: 5.1.0
argument-hint: "[command] [duration]"
---

# perf-benchmarker

Run sequential benchmarks with strict duration rules.

Follow `docs/perf-requirements.md` as the canonical contract.

## Parse Arguments

```javascript
const args = '$ARGUMENTS'.split(' ').filter(Boolean);
const command = args.find(a => !a.match(/^\d+$/)) || '';
const duration = parseInt(args.find(a => a.match(/^\d+$/)) || '60', 10);
```

## Required Rules

- Benchmarks MUST run sequentially (never parallel).
- Minimum duration: 60s per run (30s only for binary search).
- Warmup: 10s minimum before measurement.
- Re-run anomalies.

## Output Format

```
command: <benchmark command>
duration: <seconds>
warmup: <seconds>
results: <metrics summary>
notes: <anomalies or reruns>
```

## Output Contract

Benchmarks MUST emit a JSON metrics block between markers:

```
PERF_METRICS_START
{"scenarios":{"low":{"latency_ms":120},"high":{"latency_ms":450}}}
PERF_METRICS_END
```

## Constraints

- No short runs unless binary-search phase.
- Do not change code while benchmarking.
perf-benchmarker | SkillHub