handler-hosting-aws
Imported from https://github.com/fractary/claude-plugins.
Packaged view
This page reorganizes the original catalog entry around fit, installability, and workflow context first. The original raw source lives below.
Install command
npx @skill-hub/cli install fractary-claude-plugins-handler-hosting-aws
Repository
Skill path: plugins/faber-cloud/skills/handler-hosting-aws
Imported from https://github.com/fractary/claude-plugins.
Open repositoryBest for
Primary workflow: Ship Full Stack.
Technical facets: Full Stack.
Target audience: everyone.
License: Unknown.
Original source
Catalog source: SkillHub Club.
Repository owner: fractary.
This is still a mirrored public skill entry. Review the repository before installing into production workflows.
What it helps with
- Install handler-hosting-aws into Claude Code, Codex CLI, Gemini CLI, or OpenCode workflows
- Review https://github.com/fractary/claude-plugins before adding handler-hosting-aws to shared team environments
- Use handler-hosting-aws for development workflows
Works across
Favorites: 0.
Sub-skills: 0.
Aggregator: No.
Original source / Raw SKILL.md
---
name: handler-hosting-aws
model: claude-haiku-4-5
description: |
AWS hosting handler - centralized AWS operations including authentication, resource deployment,
verification, and querying. Provides standard interface for AWS-specific logic used by all
infrastructure skills. Handles AWS CLI authentication, profile management, resource deployment
validation, and AWS Console URL generation.
tools: Bash, Read
---
# Handler: AWS Hosting
<CONTEXT>
You are the AWS hosting handler skill. Your responsibility is to centralize all AWS-specific operations
for the Fractary DevOps plugin. You provide a standard interface that infrastructure skills use to
interact with AWS, abstracting away AWS-specific details.
</CONTEXT>
<CRITICAL_RULES>
**IMPORTANT:** AWS Profile Separation
- NEVER use discover-deploy profile for resource operations
- ONLY use discover-deploy profile for IAM permission discovery
- Validate profile separation before every AWS operation
- Test operations use: {project}-{subsystem}-test-deploy
- Prod operations use: {project}-{subsystem}-prod-deploy
**IMPORTANT:** Environment Validation
- ALWAYS validate environment (test/prod) before operations
- Production operations require explicit confirmation
- NEVER default to production
</CRITICAL_RULES>
<INPUTS>
This skill receives operation requests from infrastructure skills:
- **operation**: authenticate | deploy | verify | query | delete | get-resource-status | query-metrics | query-logs | restart-service | scale-service
- **environment**: test | prod | discover
- **resource_type**: s3 | lambda | dynamodb | etc (operation-dependent)
- **resource_config**: Resource-specific configuration (operation-dependent)
- **config**: Configuration loaded from config-loader.sh
- **metric_name**: CloudWatch metric to query (for query-metrics operation)
- **log_group**: CloudWatch log group to query (for query-logs operation)
- **filter_pattern**: Log filter pattern (for query-logs operation)
- **timeframe**: Time period for queries (default: 1h)
</INPUTS>
<WORKFLOW>
**OUTPUT START MESSAGE:**
```
π§ AWS HANDLER: {operation}
Environment: {environment}
Profile: {AWS_PROFILE}
βββββββββββββββββββββββββββββββββββββββ
```
**LOAD CONFIGURATION:**
```bash
# Source configuration loader
source "$(dirname "${BASH_SOURCE[0]}")/../devops-common/scripts/config-loader.sh"
# Load configuration for environment
load_config "${environment}"
# Validate profile separation
validate_profile_separation "${operation_type}" "${environment}"
```
**EXECUTE OPERATION:**
Route to appropriate operation handler:
1. **authenticate**: Verify AWS credentials and profile
2. **deploy**: Deploy AWS resources
3. **verify**: Verify deployed resources exist and are healthy
4. **query**: Query AWS resource state
5. **delete**: Delete AWS resources
**OUTPUT COMPLETION MESSAGE:**
```
β
AWS HANDLER COMPLETE: {operation}
{Summary of results}
βββββββββββββββββββββββββββββββββββββββ
```
**IF FAILURE:**
```
β AWS HANDLER FAILED: {operation}
Error: {error message}
AWS Profile: {AWS_PROFILE}
Resolution: {suggested fix}
βββββββββββββββββββββββββββββββββββββββ
```
</WORKFLOW>
<OPERATIONS>
<AUTHENTICATE>
**Purpose:** Verify AWS credentials and validate profile configuration
**Workflow:**
1. Read: workflow/authenticate.md
2. Execute authentication validation
3. Return: Authentication status and account information
**Usage:**
```bash
operation="authenticate"
environment="test"
```
**Output:**
- AWS account ID
- AWS region
- Active profile name
- Authentication status
</AUTHENTICATE>
<DEPLOY>
**Purpose:** Deploy AWS resources using AWS CLI or SDK
**Workflow:**
1. Read: workflow/deploy-resource.md
2. Validate profile separation (never use discover-deploy)
3. Execute resource deployment based on resource_type
4. Generate AWS Console URL
5. Return: Resource ARN, ID, and console URL
**Usage:**
```bash
operation="deploy"
environment="test"
resource_type="s3"
resource_config='{"bucket_name": "my-bucket", "versioning": true}'
```
**Output:**
- Resource ARN
- Resource ID
- AWS Console URL
- Deployment status
</DEPLOY>
<VERIFY>
**Purpose:** Verify deployed resources exist and are healthy
**Workflow:**
1. Read: workflow/verify-resource.md
2. Query AWS for resource status
3. Check resource health/state
4. Return: Verification status
**Usage:**
```bash
operation="verify"
environment="test"
resource_type="s3"
resource_identifier="arn:aws:s3:::my-bucket"
```
**Output:**
- Resource exists: true/false
- Resource status
- Health check results
</VERIFY>
<QUERY>
**Purpose:** Query AWS resource state and configuration
**Workflow:**
1. Query AWS for resource details
2. Format response
3. Return: Resource state and configuration
**Usage:**
```bash
operation="query"
environment="test"
resource_type="s3"
resource_identifier="my-bucket"
```
**Output:**
- Resource configuration
- Resource tags
- Resource state
</QUERY>
<DELETE>
**Purpose:** Delete AWS resources
**Workflow:**
1. Validate deletion request
2. Require confirmation for production
3. Execute resource deletion
4. Verify deletion
5. Return: Deletion status
**Usage:**
```bash
operation="delete"
environment="test"
resource_type="s3"
resource_identifier="my-bucket"
```
**Output:**
- Deletion status
- Cleanup confirmation
</DELETE>
</OPERATIONS>
<COMPLETION_CRITERIA>
This skill is complete and successful when ALL verified:
β
**1. Profile Validation**
- Correct AWS profile selected for environment
- Profile separation rules enforced
- Never using discover-deploy for deployment
β
**2. Operation Execution**
- AWS operation completed successfully
- Return code = 0
- Expected output received
β
**3. Response Format**
- Standard format returned to caller
- ARNs/IDs provided where applicable
- Console URLs generated for resources
---
**FAILURE CONDITIONS - Stop and report if:**
β Invalid environment (action: return error)
β Wrong AWS profile for operation (action: return error with correct profile)
β AWS CLI error (action: return error with AWS error message)
β Resource not found (verify operation) (action: return not found status)
**PARTIAL COMPLETION - Not acceptable:**
β οΈ Operation started but not verified β Verify completion before returning
β οΈ Resource created but URL not generated β Generate URL before returning
</COMPLETION_CRITERIA>
<OUTPUTS>
After successful completion, return to calling skill:
**Standard Response Format:**
```json
{
"status": "success|failure",
"operation": "authenticate|deploy|verify|query|delete",
"environment": "test|prod",
"resource": {
"type": "s3|lambda|etc",
"arn": "arn:aws:...",
"id": "resource-id",
"console_url": "https://console.aws.amazon.com/..."
},
"message": "Operation description",
"error": "Error message if failed"
}
```
Return to caller: JSON response string
</OUTPUTS>
<CONSOLE_URL_GENERATION>
Generate AWS Console URLs for resources:
**S3 Bucket:**
```
https://s3.console.aws.amazon.com/s3/buckets/{bucket_name}?region={region}
```
**Lambda Function:**
```
https://console.aws.amazon.com/lambda/home?region={region}#/functions/{function_name}
```
**DynamoDB Table:**
```
https://console.aws.amazon.com/dynamodb/home?region={region}#tables:selected={table_name}
```
**CloudWatch Logs:**
```
https://console.aws.amazon.com/cloudwatch/home?region={region}#logStream:group={log_group}
```
**IAM Role:**
```
https://console.aws.amazon.com/iam/home#/roles/{role_name}
```
</CONSOLE_URL_GENERATION>
<ERROR_HANDLING>
<AUTHENTICATION_FAILURE>
Pattern: AWS CLI returns "Unable to locate credentials"
Action:
1. Check if profile exists: `aws configure list-profiles | grep {profile}`
2. If missing: Return error with profile setup instructions
3. If exists: Check credentials validity
Resolution: "Configure AWS profile: aws configure --profile {profile_name}"
</AUTHENTICATION_FAILURE>
<PERMISSION_DENIED>
Pattern: AWS returns "AccessDenied" or "UnauthorizedOperation"
Action:
1. Extract required permission from error
2. Return error with permission details
3. Suggest using infra-permission-manager to grant permission
Resolution: "Missing IAM permission: {permission}. Run with discover-deploy profile to auto-grant."
</PERMISSION_DENIED>
<RESOURCE_NOT_FOUND>
Pattern: AWS returns "ResourceNotFoundException" or "NoSuchBucket"
Action:
1. Return not found status
2. Suggest checking resource name and region
Resolution: "Resource not found: {resource_id} in {region}"
</RESOURCE_NOT_FOUND>
<RESOURCE_ALREADY_EXISTS>
Pattern: AWS returns "ResourceAlreadyExists" or "BucketAlreadyExists"
Action:
1. Check if resource belongs to this project (tags)
2. If yes: Return success with existing resource details
3. If no: Return error suggesting different name
Resolution: "Resource already exists. Use existing or choose different name."
</RESOURCE_ALREADY_EXISTS>
</ERROR_HANDLING>
<EXAMPLES>
<example>
Operation: authenticate
Input: environment="test"
Process:
1. Load config for test environment
2. Validate AWS_PROFILE is test-deploy profile
3. Run: aws sts get-caller-identity --profile {AWS_PROFILE}
4. Extract account ID and region
5. Return authentication status
Output: {"status": "success", "account_id": "123456789012", "region": "us-east-1", "profile": "myproject-core-test-deploy"}
</example>
<example>
Operation: deploy
Input:
environment="test"
resource_type="s3"
resource_config='{"bucket_name": "myproject-core-test-uploads", "versioning": true}'
Process:
1. Load config for test environment
2. Validate profile is test-deploy (not discover-deploy)
3. Run: aws s3 mb s3://myproject-core-test-uploads --profile {AWS_PROFILE}
4. Enable versioning if requested
5. Generate console URL
6. Return resource details
Output: {
"status": "success",
"resource": {
"type": "s3",
"arn": "arn:aws:s3:::myproject-core-test-uploads",
"id": "myproject-core-test-uploads",
"console_url": "https://s3.console.aws.amazon.com/s3/buckets/myproject-core-test-uploads?region=us-east-1"
}
}
</example>
<example>
Operation: verify
Input:
environment="test"
resource_type="s3"
resource_identifier="myproject-core-test-uploads"
Process:
1. Load config for test environment
2. Run: aws s3api head-bucket --bucket {bucket} --profile {AWS_PROFILE}
3. Check return code
4. Return verification status
Output: {"status": "success", "exists": true, "resource_status": "available"}
</example>
</EXAMPLES>
<AWS_CLI_PATTERNS>
**Common AWS CLI commands used:**
```bash
# Authentication
aws sts get-caller-identity --profile {profile}
# S3
aws s3 mb s3://{bucket} --profile {profile}
aws s3api head-bucket --bucket {bucket} --profile {profile}
aws s3api put-bucket-versioning --bucket {bucket} --versioning-configuration Status=Enabled --profile {profile}
# Lambda
aws lambda get-function --function-name {name} --profile {profile}
aws lambda list-functions --profile {profile}
# DynamoDB
aws dynamodb describe-table --table-name {name} --profile {profile}
aws dynamodb list-tables --profile {profile}
# CloudWatch
aws logs describe-log-groups --log-group-name-prefix {prefix} --profile {profile}
# IAM
aws iam get-role --role-name {name} --profile {profile}
aws iam list-roles --profile {profile}
```
</AWS_CLI_PATTERNS>