ai-tool-assessment
Assess and report on AI tool subscriptions, usage patterns, and cost-effectiveness. Use for reviewing AI subscriptions, analyzing tool usage, optimizing AI spend.
Packaged view
This page reorganizes the original catalog entry around fit, installability, and workflow context first. The original raw source lives below.
Install command
npx @skill-hub/cli install vamseeachanta-workspace-hub-ai-tool-assessment
Repository
Skill path: .claude/skills/tools/ai-tool-assessment
Assess and report on AI tool subscriptions, usage patterns, and cost-effectiveness. Use for reviewing AI subscriptions, analyzing tool usage, optimizing AI spend.
Open repositoryBest for
Primary workflow: Analyze Data & AI.
Technical facets: Full Stack, Data / AI.
Target audience: everyone.
License: Unknown.
Original source
Catalog source: SkillHub Club.
Repository owner: vamseeachanta.
This is still a mirrored public skill entry. Review the repository before installing into production workflows.
What it helps with
- Install ai-tool-assessment into Claude Code, Codex CLI, Gemini CLI, or OpenCode workflows
- Review https://github.com/vamseeachanta/workspace-hub before adding ai-tool-assessment to shared team environments
- Use ai-tool-assessment for development workflows
Works across
Favorites: 0.
Sub-skills: 0.
Aggregator: No.
Original source / Raw SKILL.md
--- name: ai-tool-assessment description: Assess and report on AI tool subscriptions, usage patterns, and cost-effectiveness. Use for reviewing AI subscriptions, analyzing tool usage, optimizing AI spend. version: 2.0.0 category: tools last_updated: 2026-01-02 related_skills: - background-service-manager - session-start-routine --- # AI Tool Usage Assessment Assess and report on AI tool subscriptions, usage patterns, and cost-effectiveness. ## Overview This skill provides a structured framework for evaluating AI tool subscriptions, identifying underutilized services, and optimizing AI-related spending. It generates comprehensive reports with actionable recommendations. ## Quick Start 1. **Trigger assessment** - Ask to "assess AI tools" or "review AI subscriptions" 2. **Answer questions** - Provide usage patterns when asked 3. **Review report** - Generated at `reports/ai-tool-assessment/assessment-YYYYMMDD.md` 4. **Take action** - Follow recommendations for optimization ```bash # View latest assessment cat reports/ai-tool-assessment/assessment-$(date +%Y%m%d).md # List all assessments ls -la reports/ai-tool-assessment/ ``` ## When to Use - Monthly subscription reviews - Quarterly budget planning - When evaluating new AI tools - After significant workflow changes - Cost optimization initiatives - Annual AI strategy reviews ## Trigger User asks to assess AI tools, review AI subscriptions, or analyze AI tool usage. ## Instructions When triggered, perform the following assessment: ### 1. Subscription Inventory Review current paid memberships from `docs/AI_development_tools.md`: | Service | Plan | Monthly Cost | Annual Cost | |---------|------|--------------|-------------| | Claude (Anthropic) | Max Plan | $106.60 | $1,279.20 | | OpenAI | ChatGPT Plus | $21.28 | $255.36 | | Google AI | Pro | $19.99 | $239.88 | | GitHub Copilot | Pro | $8.88/mo | $106.60 | | **TOTAL** | | **$156.75** | **$1,881.04** | ### 2. Tool Usage Analysis Assess each tool category: **Primary AI Assistants:** - Claude Max: Code generation, complex reasoning, long-context tasks - OpenAI Plus: Alternative perspective, GPT-4o access, DALL-E - Google AI Pro: Gemini access, Google ecosystem integration **Development Tools:** - GitHub Copilot: Inline code completion, IDE integration - Claude-flow: Multi-agent orchestration - Factory.ai: Automated droids for CI/CD - Google Antigravity: Agent-first IDE (in evaluation) ### 3. Generate Assessment Report Create report at `reports/ai-tool-assessment/assessment-YYYYMMDD.md` with: ```markdown # AI Tool Usage Assessment - [DATE] ## Executive Summary - Total monthly spend: $X - Primary tools in active use: [list] - Tools underutilized: [list] - Recommended actions: [list] ## Subscription Status [Table of all subscriptions with renewal dates if known] ## Usage Patterns [Analysis of which tools are used for what purposes] ## Cost-Effectiveness Analysis | Tool | Cost/Month | Usage Level | Value Rating | |------|------------|-------------|--------------| | ... | ... | High/Medium/Low | 1-5 stars | ## Overlap Analysis [Identify redundant capabilities across tools] ## Recommendations 1. [Keep/Cancel/Downgrade recommendations] 2. [Usage optimization suggestions] 3. [New tools to consider] ## Next Review Date [Set quarterly review schedule] ``` ### 4. Questions to Ask User Before generating report, ask: 1. Which tools have you used most this month? 2. Are there specific tasks where one tool excels? 3. Any tools you haven't used in 30+ days? 4. New capabilities you need that current tools lack? ## Execution Checklist Before assessment: - [ ] Access to subscription documentation - [ ] User available for questions - [ ] Previous assessments reviewed (if any) During assessment: - [ ] All subscriptions inventoried - [ ] Usage patterns documented - [ ] Cost-effectiveness analyzed - [ ] Overlap identified - [ ] User questions asked After assessment: - [ ] Report generated in correct location - [ ] Recommendations prioritized - [ ] Next review date set - [ ] Source docs updated if needed ## Error Handling ### Common Issues **Error: Missing subscription data** - Cause: `docs/AI_development_tools.md` not found or outdated - Solution: Ask user for current subscription list, create/update doc **Error: Incomplete usage data** - Cause: User unsure about usage patterns - Solution: Suggest tracking usage for 2 weeks before full assessment **Error: Can't determine value** - Cause: Tool usage overlaps with others - Solution: Run overlap analysis first, identify unique value per tool **Error: Outdated pricing** - Cause: Subscription costs changed - Solution: Verify current pricing on vendor websites ## Output - Assessment report in `reports/ai-tool-assessment/` - Updated `docs/AI_development_tools.md` if status changes - Summary printed to console ## Metrics | Metric | Target | How to Measure | |--------|--------|----------------| | Cost per active tool | <$50/mo | Total spend / tools used weekly | | Utilization rate | >70% | Tools used / tools subscribed | | Value rating average | >3.5/5 | Average across all tools | | Overlap reduction | <20% | Redundant capabilities identified | | Assessment frequency | Quarterly | Days since last assessment | ## Report Templates ### Executive Summary Template ```markdown ## Executive Summary **Assessment Date:** YYYY-MM-DD **Review Period:** [Month/Quarter] **Total Monthly Spend:** $XXX.XX ### Key Findings 1. [Primary finding] 2. [Secondary finding] 3. [Tertiary finding] ### Immediate Actions Required - [ ] [Action 1] - [ ] [Action 2] ### Long-term Recommendations 1. [Recommendation with timeline] 2. [Recommendation with timeline] ``` ### Tool Comparison Template ```markdown ## Tool Comparison Matrix | Capability | Claude | OpenAI | Google | Copilot | |------------|--------|--------|--------|---------| | Code Generation | ★★★★★ | ★★★★☆ | ★★★☆☆ | ★★★★☆ | | Long Context | ★★★★★ | ★★★☆☆ | ★★★★☆ | N/A | | IDE Integration | ★★★★☆ | ★★☆☆☆ | ★★★☆☆ | ★★★★★ | | Cost Value | ★★★☆☆ | ★★★★☆ | ★★★★★ | ★★★★★ | ``` ## Best Practices 1. **Regular cadence** - Assess quarterly at minimum 2. **Track trends** - Compare assessments over time 3. **User feedback** - Include qualitative input 4. **Trial before commit** - Use free tiers to evaluate 5. **Document decisions** - Record why tools kept/cancelled 6. **Consider switching costs** - Factor in workflow disruption ## Related Skills - [background-service-manager](../background-service-manager/SKILL.md) - For running assessment scripts - [session-start-routine](../../meta/session-start-routine/SKILL.md) - Includes tool health checks --- ## Version History - **2.0.0** (2026-01-02): Upgraded to v2 template - added Quick Start, When to Use, Execution Checklist, Error Handling, Metrics sections; enhanced frontmatter with version, category, related_skills - **1.0.0** (2024-10-15): Initial release with subscription inventory, usage analysis, cost-effectiveness reporting, recommendation framework