Back to skills
SkillHub ClubAnalyze Data & AIDevOpsFull StackData / AI

pipeline

A workflow automation tool for running dlt pipelines and Temporal workflows in SignalRoom. Handles data syncing from multiple sources (Everflow, RedTrack, S3) with scheduled execution, status monitoring, and troubleshooting guides for common issues.

Packaged view

This page reorganizes the original catalog entry around fit, installability, and workflow context first. The original raw source lives below.

Stars
0
Hot score
74
Updated
March 19, 2026
Overall rating
A7.8
Composite score
4.6
Best-practice grade
S96.0

Install command

npx @skill-hub/cli install mmbianco78-signalroom-pipeline
data-pipelinetemporal-workflowetldata-syncautomation

Repository

mmbianco78/signalroom

Skill path: .claude/skills/pipeline

A workflow automation tool for running dlt pipelines and Temporal workflows in SignalRoom. Handles data syncing from multiple sources (Everflow, RedTrack, S3) with scheduled execution, status monitoring, and troubleshooting guides for common issues.

Open repository

Best for

Primary workflow: Analyze Data & AI.

Technical facets: DevOps, Full Stack, Data / AI.

Target audience: Data engineers and DevOps teams managing ETL pipelines for SignalRoom platform.

License: Unknown.

Original source

Catalog source: SkillHub Club.

Repository owner: mmbianco78.

This is still a mirrored public skill entry. Review the repository before installing into production workflows.

What it helps with

  • Install pipeline into Claude Code, Codex CLI, Gemini CLI, or OpenCode workflows
  • Review https://github.com/mmbianco78/signalroom before adding pipeline to shared team environments
  • Use pipeline for devops workflows

Works across

Claude CodeCodex CLIGemini CLIOpenCode

Favorites: 0.

Sub-skills: 0.

Aggregator: No.

Original source / Raw SKILL.md

---
name: pipeline
description: Run dlt pipelines and Temporal workflows for SignalRoom. Use when syncing data sources, triggering workflows, checking pipeline status, or debugging data ingestion issues.
---

# Pipeline Operations

## Available Sources

| Source | Description | Schedule |
|--------|-------------|----------|
| `everflow` | Affiliate conversions/revenue | Hourly 7am-11pm ET |
| `redtrack` | Ad spend tracking | Hourly 7am-11pm ET |
| `s3_exports` | CSV files from S3 (Sticky.io) | Daily 6am ET |
| `posthog` | PostHog analytics | Stubbed |
| `mautic` | Mautic contacts/campaigns | Stubbed |
| `google_sheets` | Google Sheets data | Stubbed |

**API Reference**: See `docs/API_REFERENCE.md` for live docs, auth, and request/response examples.

## Run Pipeline Directly (Local Testing)

```bash
# Activate environment
source .venv/bin/activate

# Run a specific source
python scripts/run_pipeline.py everflow
python scripts/run_pipeline.py redtrack
python scripts/run_pipeline.py s3_exports

# Dry run (see what would happen)
python scripts/run_pipeline.py everflow --dry-run
```

## Trigger via Temporal (Production)

```bash
# Trigger and return immediately
python scripts/trigger_workflow.py everflow

# Trigger and wait for completion
python scripts/trigger_workflow.py everflow -w

# Trigger with notification on success
python scripts/trigger_workflow.py everflow -w --notify
```

## Check Pipeline Status

```bash
# View worker logs (local)
make logs-worker

# View Fly.io logs (production)
fly logs

# Temporal Cloud UI
# https://cloud.temporal.io/namespaces/signalroom-713.nzg5u/workflows
```

## Active Schedules

View in Temporal Cloud UI or:

```bash
python scripts/setup_schedules.py --list
```

Current schedules:
- `hourly-sync-everflow-redtrack` - Hourly 7am-11pm ET
- `daily-sync-s3` - Daily 6am ET

## Troubleshooting

**"Unknown source" error:**
Source not registered in `src/signalroom/pipelines/runner.py`

**Credentials error:**
Check `.env` has correct API keys

**Database connection failed:**
- Verify Supabase pooler settings (port 6543)
- User must be `postgres.{project_ref}`

**Activity timeout:**
Pipeline took >30 min. Filter resources or increase timeout.