Back to skills
SkillHub ClubWrite Technical DocsFull StackData / AITech Writer

advanced-evaluation

Master LLM-as-a-Judge evaluation techniques including direct scoring, pairwise comparison, rubric generation, and bias mitigation. Use when building evaluation systems, comparing model outputs, or establishing quality standards for AI-generated content.

Packaged view

This page reorganizes the original catalog entry around fit, installability, and workflow context first. The original raw source lives below.

Stars
4
Hot score
81
Updated
March 20, 2026
Overall rating
C3.6
Composite score
3.6
Best-practice grade
A92.0

Install command

npx @skill-hub/cli install rohunvora-my-claude-skills-advanced-evaluation

Repository

rohunvora/my-claude-skills

Skill path: .claude/skills/advanced-evaluation

Master LLM-as-a-Judge evaluation techniques including direct scoring, pairwise comparison, rubric generation, and bias mitigation. Use when building evaluation systems, comparing model outputs, or establishing quality standards for AI-generated content.

Open repository

Best for

Primary workflow: Write Technical Docs.

Technical facets: Full Stack, Data / AI, Tech Writer.

Target audience: Development teams looking for install-ready agent workflows..

License: Unknown.

Original source

Catalog source: SkillHub Club.

Repository owner: rohunvora.

This is still a mirrored public skill entry. Review the repository before installing into production workflows.

What it helps with

  • Install advanced-evaluation into Claude Code, Codex CLI, Gemini CLI, or OpenCode workflows
  • Review https://github.com/rohunvora/my-claude-skills before adding advanced-evaluation to shared team environments
  • Use advanced-evaluation for development workflows

Works across

Claude CodeCodex CLIGemini CLIOpenCode

Favorites: 0.

Sub-skills: 0.

Aggregator: No.

advanced-evaluation | SkillHub