Creates specialized AI agents with optimized system prompts using the official 4-phase SOP methodology from Desktop .claude-flow, combined with evidence-based prompting techniques and Claude Agent SDK implementation. Use this skill when creating production-ready agents for specific domains, workflows, or tasks requiring consistent high-quality performance with deeply embedded domain knowledge.
Inherits all available tools
Additional assets for this skill
This skill inherits all available tools. When active, it can use any tool Claude has access to.
examples/example-1-basic.mdexamples/example-2-coordinator.mdexamples/example-3-hybrid.mdgraphviz/agent-creator-gold-process.dotgraphviz/agent-creator-process.dotgraphviz/workflow.dotreadme.mdreferences/agent-types.mdreferences/best-practices.mdreferences/index.mdreferences/integration-patterns.mdresources/readme.mdresources/scripts/4_phase_sop.pyresources/scripts/test_agent.pyresources/scripts/validate_prompt.shresources/templates/evidence-based-prompt.yamlresources/templates/system-prompt-template.mdtests/test-1-basic-agent.mdtests/test-2-complex-agent.mdtests/test-3-4phase-sop.mdname: agent-creator description: Creates specialized AI agents with optimized system prompts using the official 4-phase SOP methodology from Desktop .claude-flow, combined with evidence-based prompting techniques and Claude Agent SDK implementation. Use this skill when creating production-ready agents for specific domains, workflows, or tasks requiring consistent high-quality performance with deeply embedded domain knowledge. version: 1.0.0 category: foundry tags:
NEVER:
self_consistency: "After agent creation, test with same task multiple times to verify consistent outputs and reasoning quality" program_of_thought: "Decompose agent creation into: 1) Domain analysis, 2) Capability mapping, 3) Prompt architecture, 4) Test design, 5) Validation, 6) Integration" plan_and_solve: "Plan: Research domain + identify capabilities -> Execute: Build prompts + test cases -> Verify: Multi-run consistency + edge case handling"
<!-- END SKILL SOP IMPROVEMENT -->This skill provides the official comprehensive framework for creating specialized AI agents, integrating the proven 4-phase methodology from Desktop .claude-flow with Claude Agent SDK implementation and evidence-based prompting techniques.
Use agent-creator for:
Source: Desktop .claude-flow/ official SOP documentation
Total Time: 2.5-4 hours per agent (first-time), 1.5-2 hours (speed-run)
This methodology was developed through systematic reverse engineering of fog-compute agent creation and validated through production use.
Objective: Deep domain understanding through systematic research, not assumptions.
Activities:
Domain Breakdown
Technology Stack Mapping
Integration Points
Validation Gate:
Outputs:
Objective: Identify the cognitive expertise domains activated when you reason about this agent's tasks.
Activities:
Expertise Domain Identification
Agent Specification Creation
# Agent Specification: [Name]
## Role & Expertise
- Primary role: [Specific title]
- Expertise domains: [List activated domains]
- Cognitive patterns: [Heuristics used]
## Core Capabilities
1. [Capability with specific examples]
2. [Capability with specific examples]
...
## Decision Frameworks
- When X, do Y because Z
- Always check A before B
- Never skip validation of C
## Quality Standards
- Output must meet [criteria]
- Performance measured by [metrics]
- Failure modes to prevent: [list]
Supporting Artifacts
Validation Gate:
Outputs:
Objective: Transform specification into production-ready base system prompt.
Activities:
System Prompt Structure Design
# [AGENT NAME] - SYSTEM PROMPT v1.0
## 🎭 CORE IDENTITY
I am a **[Role Title]** with comprehensive, deeply-ingrained knowledge of [domain]. Through systematic reverse engineering and domain expertise, I possess precision-level understanding of:
- **[Domain Area 1]** - [Specific capabilities from Phase 2]
- **[Domain Area 2]** - [Specific capabilities from Phase 2]
- **[Domain Area 3]** - [Specific capabilities from Phase 2]
My purpose is to [primary objective] by leveraging [unique expertise].
## 📋 UNIVERSAL COMMANDS I USE
**File Operations**:
- /file-read, /file-write, /glob-search, /grep-search
WHEN: [Specific situations from domain analysis]
HOW: [Exact patterns]
**Git Operations**:
- /git-status, /git-commit, /git-push
WHEN: [Specific situations]
HOW: [Exact patterns]
**Communication & Coordination**:
- /memory-store, /memory-retrieve
- /agent-delegate, /agent-escalate
WHEN: [Specific situations]
HOW: [Exact patterns with namespace conventions]
## 🎯 MY SPECIALIST COMMANDS
[List role-specific commands with exact syntax and examples]
## 🔧 MCP SERVER TOOLS I USE
**Claude Flow MCP**:
- mcp__claude-flow__agent_spawn
WHEN: [Specific coordination scenarios]
HOW: [Exact function call patterns]
- mcp__claude-flow__memory_store
WHEN: [Cross-agent data sharing]
HOW: [Namespace pattern: agent-role/task-id/data-type]
**[Other relevant MCP servers from Phase 1]**
## 🧠 COGNITIVE FRAMEWORK
### Self-Consistency Validation
Before finalizing deliverables, I validate from multiple angles:
1. [Domain-specific validation 1]
2. [Domain-specific validation 2]
3. [Cross-check with standards]
### Program-of-Thought Decomposition
For complex tasks, I decompose BEFORE execution:
1. [Domain-specific decomposition pattern]
2. [Dependency analysis]
3. [Risk assessment]
### Plan-and-Solve Execution
My standard workflow:
1. PLAN: [Domain-specific planning]
2. VALIDATE: [Domain-specific validation]
3. EXECUTE: [Domain-specific execution]
4. VERIFY: [Domain-specific verification]
5. DOCUMENT: [Memory storage patterns]
## 🚧 GUARDRAILS - WHAT I NEVER DO
[From Phase 2 failure modes and edge cases]
**[Failure Category 1]**:
❌ NEVER: [Dangerous pattern]
WHY: [Consequences from domain knowledge]
WRONG:
[Bad example]
CORRECT:
[Good example]
## ✅ SUCCESS CRITERIA
Task complete when:
- [ ] [Domain-specific criterion 1]
- [ ] [Domain-specific criterion 2]
- [ ] [Domain-specific criterion 3]
- [ ] Results stored in memory
- [ ] Relevant agents notified
## 📖 WORKFLOW EXAMPLES
### Workflow 1: [Common Task Name from Phase 1]
**Objective**: [What this achieves]
**Step-by-Step Commands**:
```yaml
Step 1: [Action]
COMMANDS:
- /[command-1] --params
- /[command-2] --params
OUTPUT: [Expected]
VALIDATION: [Check]
Step 2: [Next Action]
COMMANDS:
- /[command-3] --params
OUTPUT: [Expected]
VALIDATION: [Check]
Timeline: [Duration] Dependencies: [Prerequisites]
Evidence-Based Technique Integration
For each technique (from existing agent-creator skill):
Integrate these naturally into the agent's methodology.
Quality Standards & Guardrails
From Phase 2 failure modes, create explicit guardrails:
Validation Gate:
Outputs:
Objective: Reverse-engineer exact implementation patterns and document with precision.
Activities:
Code Pattern Extraction
For technical agents, extract EXACT patterns from codebase:
## Code Patterns I Recognize
### Pattern: [Name]
**File**: `path/to/file.py:123-156`
```python
class ExamplePattern:
def __init__(
self,
param1: Type = default, # Line 125: Exact default
param2: Type = default # Line 126: Exact default
):
# Extracted from actual implementation
pass
When I see this pattern, I know:
Critical Failure Mode Documentation
From experience and domain knowledge:
## Critical Failure Modes
### Failure: [Name]
**Severity**: Critical/High/Medium
**Symptoms**: [How to recognize]
**Root Cause**: [Why it happens]
**Prevention**:
❌ DON'T: [Bad pattern]
✅ DO: [Good pattern with exact code]
**Detection**:
```bash
# Exact command to detect this failure
[command]
Integration Patterns
Document exact MCP tool usage:
## MCP Integration Patterns
### Pattern: Cross-Agent Data Sharing
```javascript
// Exact pattern for storing outputs
mcp__claude-flow__memory_store({
key: "marketing-specialist/campaign-123/audience-analysis",
value: {
segments: [...],
targeting: {...},
confidence: 0.89
},
ttl: 86400
})
Namespace Convention:
{agent-role}/{task-id}/{data-type}backend-dev/api-v2/schema-design
Performance Metrics
Define what to track:
## Performance Metrics I Track
```yaml
Task Completion:
- /memory-store --key "metrics/[my-role]/tasks-completed" --increment 1
- /memory-store --key "metrics/[my-role]/task-[id]/duration" --value [ms]
Quality:
- validation-passes: [count successful validations]
- escalations: [count when needed help]
- error-rate: [failures / attempts]
Efficiency:
- commands-per-task: [avg commands used]
- mcp-calls: [tool usage frequency]
These metrics enable continuous improvement.
Validation Gate:
Outputs:
Combining 4-phase SOP with existing best practices:
Phase 1: Domain Analysis (30-60 min)
Phase 2: Expertise Extraction (30-45 min)
Phase 3: Architecture Design (45-60 min)
Phase 4: Technical Enhancement (60-90 min)
SDK Implementation (30-60 min)
Testing & Validation (30-45 min)
Documentation & Packaging (15-30 min)
Total Time: 3.5-5.5 hours (first-time), 2-3 hours (speed-run)
Once system prompt is finalized, implement with SDK:
import { query, tool } from '@anthropic-ai/claude-agent-sdk';
import { z } from 'zod';
// Custom domain-specific tools
const domainTool = tool({
name: 'domain_operation',
description: 'Performs domain-specific operation',
parameters: z.object({
param: z.string()
}),
handler: async ({ param }) => {
// Implementation from Phase 4
return { result: 'data' };
}
});
// Agent configuration
for await (const message of query('Perform domain task', {
model: 'claude-sonnet-4-5',
systemPrompt: enhancedPromptV2, // From Phase 4
permissionMode: 'acceptEdits',
allowedTools: ['Read', 'Write', 'Bash', domainTool],
mcpServers: [{
command: 'npx',
args: ['claude-flow@alpha', 'mcp', 'start'],
env: { ... }
}],
settingSources: ['user', 'project']
})) {
console.log(message);
}
from claude_agent_sdk import query, tool, ClaudeAgentOptions
import asyncio
@tool()
async def domain_operation(param: str) -> dict:
"""Domain-specific operation from Phase 4."""
# Implementation
return {"result": "data"}
async def run_agent():
options = ClaudeAgentOptions(
model='claude-sonnet-4-5',
system_prompt=enhanced_prompt_v2, # From Phase 4
permission_mode='acceptEdits',
allowed_tools=['Read', 'Write', 'Bash', domain_operation],
mcp_servers=[{
'command': 'npx',
'args': ['claude-flow@alpha', 'mcp', 'start']
}],
setting_sources=['user', 'project']
)
async for message in query('Perform domain task', **options):
print(message)
asyncio.run(run_agent())
From existing agent-creator skill, enhanced with 4-phase methodology:
Phase 1 Focus: Evidence evaluation patterns, data quality standards Phase 2 Focus: Analytical heuristics, validation frameworks Phase 3 Focus: Self-consistency checking, confidence calibration Phase 4 Focus: Statistical validation code, error detection patterns
Phase 1 Focus: Quality criteria, template patterns Phase 2 Focus: Creative heuristics, refinement cycles Phase 3 Focus: Plan-and-solve frameworks, requirement tracking Phase 4 Focus: Generation patterns, quality validation code
Phase 1 Focus: Problem patterns, debugging workflows Phase 2 Focus: Hypothesis generation, systematic testing Phase 3 Focus: Program-of-thought decomposition, evidence tracking Phase 4 Focus: Detection scripts, root cause analysis patterns
Phase 1 Focus: Workflow patterns, dependency management Phase 2 Focus: Coordination heuristics, error recovery Phase 3 Focus: Plan-and-solve with dependencies, progress tracking Phase 4 Focus: Orchestration code, retry logic, escalation paths
From existing framework + SOP enhancements:
Phase 1 (Analysis):
Phase 2 (Expertise Extraction):
Phase 3 (Architecture):
Phase 4 (Enhancement):
Total: 2 hours for experienced creators with templates
See: docs/agent-architecture/agents-rewritten/MARKETING-SPECIALIST-AGENT.md
Phase 1 Output: Marketing domain analysis, tools (Google Analytics, SEMrush, etc.) Phase 2 Output: Marketing expertise (CAC, LTV, funnel optimization, attribution) Phase 3 Output: Base prompt with 9 specialist commands Phase 4 Output: Campaign workflow patterns, A/B test validation, ROI calculations
Result: Production-ready agent with deeply embedded marketing expertise
This enhanced agent-creator skill combines:
Use this methodology to create all 90 specialist agents with:
Next: Begin agent rewrites using this enhanced methodology.