Back to skills
SkillHub ClubBuild MobileFull StackMobile

axiom-ios-vision

Use when implementing ANY computer vision feature - image analysis, object detection, pose detection, person segmentation, subject lifting, hand/body pose tracking.

Packaged view

This page reorganizes the original catalog entry around fit, installability, and workflow context first. The original raw source lives below.

Stars
657
Hot score
99
Updated
March 20, 2026
Overall rating
C4.2
Composite score
4.2
Best-practice grade
A92.4

Install command

npx @skill-hub/cli install charleswiltgen-axiom-axiom-ios-vision

Repository

CharlesWiltgen/Axiom

Skill path: .claude-plugin/plugins/axiom/skills/axiom-ios-vision

Use when implementing ANY computer vision feature - image analysis, object detection, pose detection, person segmentation, subject lifting, hand/body pose tracking.

Open repository

Best for

Primary workflow: Build Mobile.

Technical facets: Full Stack, Mobile.

Target audience: everyone.

License: Unknown.

Original source

Catalog source: SkillHub Club.

Repository owner: CharlesWiltgen.

This is still a mirrored public skill entry. Review the repository before installing into production workflows.

What it helps with

  • Install axiom-ios-vision into Claude Code, Codex CLI, Gemini CLI, or OpenCode workflows
  • Review https://github.com/CharlesWiltgen/Axiom before adding axiom-ios-vision to shared team environments
  • Use axiom-ios-vision for development workflows

Works across

Claude CodeCodex CLIGemini CLIOpenCode

Favorites: 0.

Sub-skills: 0.

Aggregator: No.

Original source / Raw SKILL.md

---
name: axiom-ios-vision
description: Use when implementing ANY computer vision feature - image analysis, object detection, pose detection, person segmentation, subject lifting, hand/body pose tracking.
user-invocable: false
---

# iOS Computer Vision Router

**You MUST use this skill for ANY computer vision work using the Vision framework.**

## When to Use

Use this router when:
- Analyzing images or video
- Detecting objects, faces, or people
- Tracking hand or body pose
- Segmenting people or subjects
- Lifting subjects from backgrounds
- Recognizing text in images (OCR)
- Detecting barcodes or QR codes
- Scanning documents
- Using VisionKit or DataScannerViewController

## Routing Logic

### Vision Work

**Implementation patterns** → `/skill axiom-vision`
- Subject segmentation (VisionKit)
- Hand pose detection (21 landmarks)
- Body pose detection (2D/3D)
- Person segmentation
- Face detection
- Isolating objects while excluding hands
- Text recognition (VNRecognizeTextRequest)
- Barcode/QR detection (VNDetectBarcodesRequest)
- Document scanning (VNDocumentCameraViewController)
- Live scanning (DataScannerViewController)
- Structured document extraction (RecognizeDocumentsRequest, iOS 26+)

**API reference** → `/skill axiom-vision-ref`
- Complete Vision framework API
- VNDetectHumanHandPoseRequest
- VNDetectHumanBodyPoseRequest
- VNGenerateForegroundInstanceMaskRequest
- VNRecognizeTextRequest (fast/accurate modes)
- VNDetectBarcodesRequest (symbologies)
- DataScannerViewController delegates
- RecognizeDocumentsRequest (iOS 26+)
- Coordinate conversion patterns

**Diagnostics** → `/skill axiom-vision-diag`
- Subject not detected
- Hand pose missing landmarks
- Low confidence observations
- Performance issues
- Coordinate conversion bugs
- Text not recognized or wrong characters
- Barcodes not detected
- DataScanner showing blank or no items
- Document edges not detected

## Decision Tree

```
User asks about computer vision
  ├─ Implementing?
  │   ├─ Pose detection (hand/body)? → vision
  │   ├─ Subject segmentation? → vision
  │   ├─ Text recognition/OCR? → vision
  │   ├─ Barcode/QR scanning? → vision
  │   ├─ Document scanning? → vision
  │   └─ Live camera scanning? → vision (DataScannerViewController)
  ├─ Need API reference? → vision-ref
  └─ Debugging issues? → vision-diag
```

## Critical Patterns

**vision**:
- Subject segmentation with VisionKit
- Hand pose detection (21 landmarks)
- Body pose detection (2D/3D, up to 4 people)
- Isolating objects while excluding hands
- CoreImage HDR compositing
- Text recognition (fast vs accurate modes)
- Barcode detection (symbology selection)
- Document scanning with perspective correction
- Live scanning with DataScannerViewController
- Structured document extraction (iOS 26+)

**vision-diag**:
- Subject detection failures
- Landmark tracking issues
- Performance optimization
- Observation confidence thresholds
- Text recognition failures (language, contrast)
- Barcode detection issues (symbology, distance)
- DataScanner troubleshooting
- Document edge detection problems

## Example Invocations

User: "How do I detect hand pose in an image?"
→ Invoke: `/skill axiom-vision`

User: "Isolate a subject but exclude the user's hands"
→ Invoke: `/skill axiom-vision`

User: "How do I read text from an image?"
→ Invoke: `/skill axiom-vision`

User: "Scan QR codes with the camera"
→ Invoke: `/skill axiom-vision`

User: "How do I implement document scanning?"
→ Invoke: `/skill axiom-vision`

User: "Use DataScannerViewController for live text"
→ Invoke: `/skill axiom-vision`

User: "Subject detection isn't working"
→ Invoke: `/skill axiom-vision-diag`

User: "Text recognition returns wrong characters"
→ Invoke: `/skill axiom-vision-diag`

User: "Barcode not being detected"
→ Invoke: `/skill axiom-vision-diag`

User: "Show me VNDetectHumanBodyPoseRequest examples"
→ Invoke: `/skill axiom-vision-ref`

User: "What symbologies does VNDetectBarcodesRequest support?"
→ Invoke: `/skill axiom-vision-ref`

User: "RecognizeDocumentsRequest API reference"
→ Invoke: `/skill axiom-vision-ref`
axiom-ios-vision | SkillHub