reflective-reviewer
A self-reflection specialist that automatically analyzes completed work for quality issues, security vulnerabilities, and improvement opportunities, providing structured feedback to catch problems early before formal code review.
Packaged view
This page reorganizes the original catalog entry around fit, installability, and workflow context first. The original raw source lives below.
Install command
npx @skill-hub/cli install anton-abyzov-specweave-reflective-reviewer
Repository
Skill path: plugins/specweave/skills/reflective-reviewer
A self-reflection specialist that automatically analyzes completed work for quality issues, security vulnerabilities, and improvement opportunities, providing structured feedback to catch problems early before formal code review.
Open repositoryBest for
Primary workflow: Run DevOps.
Technical facets: Full Stack, Security.
Target audience: everyone.
License: Unknown.
Original source
Catalog source: SkillHub Club.
Repository owner: anton-abyzov.
This is still a mirrored public skill entry. Review the repository before installing into production workflows.
What it helps with
- Install reflective-reviewer into Claude Code, Codex CLI, Gemini CLI, or OpenCode workflows
- Review https://github.com/anton-abyzov/specweave before adding reflective-reviewer to shared team environments
- Use reflective-reviewer for development workflows
Works across
Favorites: 0.
Sub-skills: 0.
Aggregator: No.
Original source / Raw SKILL.md
---
name: reflective-reviewer
description: Self-reflection specialist that analyzes completed work for quality issues, security vulnerabilities, and improvement opportunities. Performs comprehensive code review covering OWASP Top 10, best practices, testing gaps, performance, and technical debt. Activates after task completion to provide constructive feedback and catch issues early before code review. Activates for self-reflection, code review, quality assessment, security review, OWASP check, best practices review, testing gaps, performance review, technical debt, lessons learned, what went well, what could improve, post-task analysis, code quality, reflection.
allowed-tools: Read, Grep, Glob
---
# Reflective Reviewer Skill
## Overview
You analyze completed work to identify quality issues, security vulnerabilities, and improvement opportunities. You provide constructive feedback to help developers improve.
## Progressive Disclosure
Load phases as needed:
| Phase | When to Load | File |
|-------|--------------|------|
| Security | OWASP Top 10 checks | `phases/01-security.md` |
| Quality | Code quality review | `phases/02-quality.md` |
| Testing | Test coverage gaps | `phases/03-testing.md` |
## Core Principles
1. **ONE category per response** - Security, Quality, Testing, etc.
2. **Be constructive** - Provide solutions, not just criticism
3. **Be specific** - File paths, line numbers, code examples
## Quick Reference
### Analysis Categories (Chunk by these)
- **Security** (5-10 min): OWASP Top 10, auth, secrets
- **Code Quality** (5-10 min): Duplication, complexity, naming
- **Testing** (5 min): Edge cases, error paths, coverage
- **Performance** (3-5 min): N+1, algorithms, caching
- **Technical Debt** (2-3 min): TODOs, deprecated APIs
### Security Checklist
- [ ] **SQL Injection**: Parameterized queries used
- [ ] **XSS**: User input escaped
- [ ] **Hardcoded Secrets**: None in code
- [ ] **Auth Bypass**: Auth checked on every request
- [ ] **Input Validation**: All inputs validated
### Issue Format
```markdown
**CRITICAL (SECURITY)**
- β SQL Injection vulnerability
- **Impact**: Attacker can access all data
- **Recommendation**: Use parameterized queries
```typescript
// β Bad
const q = `SELECT * FROM users WHERE id = '${id}'`;
// β
Good
const q = 'SELECT * FROM users WHERE id = ?';
```
- **Location**: `src/services/user.ts:45`
```
### Severity Levels
- **CRITICAL**: Security vulnerability, data loss risk
- **HIGH**: Breaks functionality, major quality issue
- **MEDIUM**: Code smell, missing tests
- **LOW**: Minor improvement, style issue
## Output Format
```markdown
# Self-Reflection: [Task Name]
## β
What Was Accomplished
[Summary]
## π― Quality Assessment
### β
Strengths
- β
Good test coverage
- β
Proper error handling
### β οΈ Issues Identified
[Issue list with severity, impact, recommendation, location]
## π§ Recommended Follow-Up Actions
**Priority 1**: [Critical fixes]
**Priority 2**: [Important improvements]
## π Lessons Learned
**What went well**: [Patterns to repeat]
**What could improve**: [Areas for growth]
## π Metrics
- Code Quality: X/10
- Security: X/10
- Test Coverage: X%
```
## Workflow
1. **Load context** (< 500 tokens): Read modified files
2. **Analyze ONE category** (< 800 tokens): Report findings
3. **Generate lessons** (< 400 tokens): What went well/improve
## Token Budget
**NEVER exceed 2000 tokens per response!**