Back to skills
SkillHub ClubRun DevOpsFull StackDevOps

Use llamactl - a CLI tool for LlamaAgents

Use llamactl to initialize, locally preview, deploy and manage LlamaIndex workflows as LlamaAgents. Required llama-index-workflows and llamactl to be installed in the environment.

Packaged view

This page reorganizes the original catalog entry around fit, installability, and workflow context first. The original raw source lives below.

Stars
174
Hot score
96
Updated
March 20, 2026
Overall rating
C3.5
Composite score
3.5
Best-practice grade
C64.0

Install command

npx @skill-hub/cli install run-llama-vibe-llama-llamactl

Repository

run-llama/vibe-llama

Skill path: documentation/skills/llamactl

Use llamactl to initialize, locally preview, deploy and manage LlamaIndex workflows as LlamaAgents. Required llama-index-workflows and llamactl to be installed in the environment.

Open repository

Best for

Primary workflow: Run DevOps.

Technical facets: Full Stack, DevOps.

Target audience: everyone.

License: Unknown.

Original source

Catalog source: SkillHub Club.

Repository owner: run-llama.

This is still a mirrored public skill entry. Review the repository before installing into production workflows.

What it helps with

  • Install Use llamactl - a CLI tool for LlamaAgents into Claude Code, Codex CLI, Gemini CLI, or OpenCode workflows
  • Review https://github.com/run-llama/vibe-llama before adding Use llamactl - a CLI tool for LlamaAgents to shared team environments
  • Use Use llamactl - a CLI tool for LlamaAgents for development workflows

Works across

Claude CodeCodex CLIGemini CLIOpenCode

Favorites: 0.

Sub-skills: 0.

Aggregator: No.

Original source / Raw SKILL.md

---
name: Use llamactl - a CLI tool for LlamaAgents
description: Use llamactl to initialize, locally preview, deploy and manage LlamaIndex workflows as LlamaAgents. Required llama-index-workflows and llamactl to be installed in the environment.
---

# Use llamactl - a CLI tool for LlamaAgents

`llamactl` is a CLI tool for developing and deploying LlamaIndex workflows as LlamaAgents. It provides commands to initialize projects, run local development servers, and manage cloud deployments.

## Prerequisites

Before using `llamactl`, ensure you have:

- [`uv`](https://docs.astral.sh/uv/getting-started/installation/) - Python package manager and build tool
- Node.js - Required for UI development (supports `npm`, `pnpm`, or `yarn`)
- `llama-index-workflows` and `llamactl` installed in your environment

## Installation

Install `llamactl` globally using `uv`:

```bash
uv tool install -U llamactl
llamactl --help
```

Or try it without installing:

```bash
uvx llamactl --help
```

## Initialize a Project

Create a new LlamaAgents project with starter templates:

```bash
llamactl init
```

This creates a Python module with LlamaIndex workflows and an optional UI frontend. Configuration is managed in `pyproject.toml`, where you define workflow instances, environment settings, and UI build options.

## Local Development

Start the local development server:

```bash
llamactl serve
```

This command:

1. Installs dependencies
2. Serves workflows as an API (configured in `pyproject.toml`)
3. Starts the frontend development server

The server automatically detects file changes and can resume in-progress workflows.

## Deploy to LlamaCloud

Push your code to a git repository:

```bash
git remote add origin https://github.com/org/repo
git add -A
git commit -m 'Set up new app'
git push -u origin main
```

Create a cloud deployment:

```bash
llamactl deployments create
```

This opens an interactive Terminal UI to configure:

- Deployment name
- Git repository (supports private GitHub repos via the LlamaDeploy GitHub app)
- Git branch/tag/commit
- Environment secrets

## Manage Deployments

- View deployment status: `llamactl deployments get`
- Update secrets or branch: `llamactl deployments edit`
- Deploy new version: `llamactl deployments update`

For detailed configuration options, see the [Deployment Config Reference](https://developers.llamaindex.ai/python/cloud/llamaagents/configuration-reference).