ai-llm-engineering
Operational skill hub for LLM system architecture, evaluation, deployment, and optimization (modern production standards). Links to specialized skills for prompts, RAG, agents, and safety. Integrates recent advances: PEFT/LoRA fine-tuning, hybrid RAG handoff (see dedicated skill), vLLM 24x throughput, multi-layered security (90%+ bypass for single-layer), automated drift detection (18-second response), and CI/CD-aligned evaluation.
Packaged view
This page reorganizes the original catalog entry around fit, installability, and workflow context first. The original raw source lives below.
Install command
npx @skill-hub/cli install microck-ordinary-claude-skills-ai-llm-engineering
Repository
Skill path: skills_categorized/wellness-health/ai-llm-engineering
Operational skill hub for LLM system architecture, evaluation, deployment, and optimization (modern production standards). Links to specialized skills for prompts, RAG, agents, and safety. Integrates recent advances: PEFT/LoRA fine-tuning, hybrid RAG handoff (see dedicated skill), vLLM 24x throughput, multi-layered security (90%+ bypass for single-layer), automated drift detection (18-second response), and CI/CD-aligned evaluation.
Open repositoryBest for
Primary workflow: Analyze Data & AI.
Technical facets: Full Stack, DevOps, Data / AI, Security.
Target audience: Development teams looking for install-ready agent workflows..
License: Unknown.
Original source
Catalog source: SkillHub Club.
Repository owner: Microck.
This is still a mirrored public skill entry. Review the repository before installing into production workflows.
What it helps with
- Install ai-llm-engineering into Claude Code, Codex CLI, Gemini CLI, or OpenCode workflows
- Review https://github.com/Microck/ordinary-claude-skills before adding ai-llm-engineering to shared team environments
- Use ai-llm-engineering for development workflows
Works across
Favorites: 0.
Sub-skills: 0.
Aggregator: No.