llm-models
Access Claude, Gemini, Kimi, GLM and 100+ LLMs via inference.sh CLI using OpenRouter. Models: Claude Opus 4.5, Claude Sonnet 4.5, Claude Haiku 4.5, Gemini 3 Pro, Kimi K2, GLM-4.6, Intellect 3. One API for all models with automatic fallback and cost optimization. Use for: AI assistants, code generation, reasoning, agents, chat, content generation. Triggers: claude api, openrouter, llm api, claude sonnet, claude opus, gemini api, kimi, language model, gpt alternative, anthropic api, ai model api, llm access, chat api, claude alternative, openai alternative
Packaged view
This page reorganizes the original catalog entry around fit, installability, and workflow context first. The original raw source lives below.
Install command
npx @skill-hub/cli install inferencesh-skills-llm-models
Repository
Skill path: skills/llm-models
Access Claude, Gemini, Kimi, GLM and 100+ LLMs via inference.sh CLI using OpenRouter. Models: Claude Opus 4.5, Claude Sonnet 4.5, Claude Haiku 4.5, Gemini 3 Pro, Kimi K2, GLM-4.6, Intellect 3. One API for all models with automatic fallback and cost optimization. Use for: AI assistants, code generation, reasoning, agents, chat, content generation. Triggers: claude api, openrouter, llm api, claude sonnet, claude opus, gemini api, kimi, language model, gpt alternative, anthropic api, ai model api, llm access, chat api, claude alternative, openai alternative
Open repositoryBest for
Primary workflow: Write Technical Docs.
Technical facets: Full Stack, Backend, Data / AI, Tech Writer.
Target audience: everyone.
License: Unknown.
Original source
Catalog source: SkillHub Club.
Repository owner: inferencesh.
This is still a mirrored public skill entry. Review the repository before installing into production workflows.
What it helps with
- Install llm-models into Claude Code, Codex CLI, Gemini CLI, or OpenCode workflows
- Review https://github.com/inferencesh/skills before adding llm-models to shared team environments
- Use llm-models for development workflows
Works across
Favorites: 0.
Sub-skills: 0.
Aggregator: No.
Original source / Raw SKILL.md
---
name: llm-models
description: |
Access Claude, Gemini, Kimi, GLM and 100+ LLMs via inference.sh CLI using OpenRouter.
Models: Claude Opus 4.5, Claude Sonnet 4.5, Claude Haiku 4.5, Gemini 3 Pro, Kimi K2, GLM-4.6, Intellect 3.
One API for all models with automatic fallback and cost optimization.
Use for: AI assistants, code generation, reasoning, agents, chat, content generation.
Triggers: claude api, openrouter, llm api, claude sonnet, claude opus, gemini api, kimi,
language model, gpt alternative, anthropic api, ai model api, llm access, chat api,
claude alternative, openai alternative
allowed-tools: Bash(infsh *)
---
# LLM Models via OpenRouter

Access 100+ language models via [inference.sh](https://inference.sh) CLI.
## Quick Start
```bash
curl -fsSL https://cli.inference.sh | sh && infsh login
# Call Claude Sonnet
infsh app run openrouter/claude-sonnet-45 --input '{"prompt": "Explain quantum computing"}'
```
## Available Models
| Model | App ID | Best For |
|-------|--------|----------|
| Claude Opus 4.5 | `openrouter/claude-opus-45` | Complex reasoning, coding |
| Claude Sonnet 4.5 | `openrouter/claude-sonnet-45` | Balanced performance |
| Claude Haiku 4.5 | `openrouter/claude-haiku-45` | Fast, economical |
| Gemini 3 Pro | `openrouter/gemini-3-pro-preview` | Google's latest |
| Kimi K2 Thinking | `openrouter/kimi-k2-thinking` | Multi-step reasoning |
| GLM-4.6 | `openrouter/glm-46` | Open-source, coding |
| Intellect 3 | `openrouter/intellect-3` | General purpose |
| Any Model | `openrouter/any-model` | Auto-selects best option |
## Search LLM Apps
```bash
infsh app list --search "openrouter"
infsh app list --search "claude"
```
## Examples
### Claude Opus (Best Quality)
```bash
infsh app run openrouter/claude-opus-45 --input '{
"prompt": "Write a Python function to detect palindromes with comprehensive tests"
}'
```
### Claude Sonnet (Balanced)
```bash
infsh app run openrouter/claude-sonnet-45 --input '{
"prompt": "Summarize the key concepts of machine learning"
}'
```
### Claude Haiku (Fast & Cheap)
```bash
infsh app run openrouter/claude-haiku-45 --input '{
"prompt": "Translate this to French: Hello, how are you?"
}'
```
### Kimi K2 (Thinking Agent)
```bash
infsh app run openrouter/kimi-k2-thinking --input '{
"prompt": "Plan a step-by-step approach to build a web scraper"
}'
```
### Any Model (Auto-Select)
```bash
# Automatically picks the most cost-effective model
infsh app run openrouter/any-model --input '{
"prompt": "What is the capital of France?"
}'
```
### With System Prompt
```bash
infsh app sample openrouter/claude-sonnet-45 --save input.json
# Edit input.json:
# {
# "system": "You are a helpful coding assistant",
# "prompt": "How do I read a file in Python?"
# }
infsh app run openrouter/claude-sonnet-45 --input input.json
```
## Use Cases
- **Coding**: Generate, review, debug code
- **Writing**: Content, summaries, translations
- **Analysis**: Data interpretation, research
- **Agents**: Build AI-powered workflows
- **Chat**: Conversational interfaces
## Related Skills
```bash
# Full platform skill (all 150+ apps)
npx skills add inferencesh/skills@inference-sh
# Web search (combine with LLMs for RAG)
npx skills add inferencesh/skills@web-search
# Image generation
npx skills add inferencesh/skills@ai-image-generation
# Video generation
npx skills add inferencesh/skills@ai-video-generation
```
Browse all apps: `infsh app list`
## Documentation
- [Agents Overview](https://inference.sh/docs/concepts/agents) - Building AI agents
- [Agent SDK](https://inference.sh/docs/api/agent/overview) - Programmatic agent control
- [Building a Research Agent](https://inference.sh/blog/guides/research-agent) - LLM + search integration guide