agent-architect

Research-backed creation and enhancement of opencode agents. Use when creating new agents from requirements, improving underperforming agents, or designing agent configurations. Triggers on "create an agent for X", "improve my Y agent", "design a specialized agent", "my agent isn't working well".

$ Installer

git clone https://github.com/IgorWarzocha/Opencode-Workflows /tmp/Opencode-Workflows && cp -r /tmp/Opencode-Workflows/opencode-configurator/skill/agent-architect ~/.claude/skills/Opencode-Workflows

// tip: Run this command in your terminal to install the skill


name: agent-architect description: Research-backed creation and enhancement of opencode agents. Use when creating new agents from requirements, improving underperforming agents, or designing agent configurations. Triggers on "create an agent for X", "improve my Y agent", "design a specialized agent", "my agent isn't working well".

Agent Architect

Create and refine opencode agents through a guided Q&A process.

<core_approach>

Agent creation is conversational, not transactional.

  • MUST NOT assume what the user wants—ask
  • SHOULD start with broad questions, drill into details only if needed
  • Users MAY skip configuration they don't care about
  • MUST always show drafts and iterate based on feedback

The goal is to help users create agents that fit their needs, not to dump every possible configuration option on them.

</core_approach>

Agent Locations

ScopePath
Project.opencode/agent/<name>.md
Global~/.config/opencode/agent/<name>.md

Agent File Format

---
description: When to use this agent. Include trigger examples.
model: anthropic/claude-sonnet-4-20250514  # Optional
mode: primary | subagent | all             # Default: all
permission:
  skill: { "my-skill": "allow", "*": "deny" }
  bash: { "git *": "allow", "*": "ask" }
---
System prompt in markdown body (second person).

Full schema: See references/opencode-config.md

Agent Modes

ModeDescription
primaryUser-selectable, can be default agent
subagentOnly callable via task tool
allBoth primary and subagent (default)

Phase 1: Core Purpose (Required)

Ask these first—they shape everything else:

  1. "What should this agent do?"

    • Get the core task/domain
    • Examples: "review code", "help with deployments", "research topics"
  2. "What should trigger this agent?"

    • Specific phrases, contexts, file types
    • Becomes the description field
  3. "What expertise/persona should it have?"

    • Tone, boundaries, specialization
    • Shapes the system prompt

Phase 1.5: Research the Domain

MUST NOT assume knowledge is current. After understanding the broad strokes:

  • Search for current best practices in the domain
  • Check for updates to frameworks, tools, or APIs the agent will work with
  • Look up documentation for any unfamiliar technologies mentioned
  • Find examples of how experts approach similar tasks

This research informs better questions in Phase 2 and produces a more capable agent.

Example: User wants an agent for "Next.js deployments" → Research current Next.js deployment patterns, Vercel vs self-hosted, App Router vs Pages Router, common pitfalls, etc.

Phase 2: Capabilities (Ask broadly, then drill down)

  1. "Does this agent need shell access? Web access? File editing?"

    • If yes to shell: "Any commands that should be blocked or require approval?"
    • If yes to web: "Should it ask before fetching URLs?"
    • Determines permission config
  2. "Should this agent use any skills?"

    • If yes: "Which ones? Should others be blocked?"
    • Configures permission.skill
  3. "Is this a primary agent or a helper for other agents?"

    • Primary = user-selectable
    • Subagent = only called via task tool
    • Sets mode

Phase 3: Details (Optional—user MAY skip)

  1. "Any specific model preference?" (most users skip)
  2. "Custom temperature/sampling?" (most users skip)
  3. "Maximum steps before stopping?" (most users skip)

Phase 4: Review & Refine

  1. Show the draft config and prompt, ask for feedback
    • "Here's what I've created. Anything you'd like to change?"
    • Iterate until user is satisfied

Key principle: Start broad, get specific only where the user shows interest. MUST NOT overwhelm with options like top_p unless asked.

Be flexible: If the user provides lots of info upfront, adapt—MUST NOT rigidly follow the phases. If they say "I want a code review agent that can't run shell commands", you already have answers to multiple questions.

<system_prompt_structure>

Recommended Structure

# Role and Objective
You are [expert persona]. Your goal is [objective].

# Instructions
- Core behavioral rules
- What to always/never do

## Sub-instructions (optional)
More detailed guidance for specific areas.

# Workflow
1. First, [step]
2. Then, [step]
3. Finally, [step]

# Output Format
Specify exact format expected.

# Examples (optional)
<examples>
<example>
<input>User request</input>
<output>Expected response</output>
</example>
</examples>

XML Tags (Recommended)

XML tags improve clarity and parseability across all models:

TagPurpose
<instructions>Core behavioral rules
<context>Background information
<examples>Few-shot demonstrations
<thinking>Chain-of-thought reasoning
<output>Final response format

Best practices:

  • Be consistent with tag names throughout
  • Nest tags for hierarchy: <outer><inner></inner></outer>
  • Reference tags in instructions: "Using the data in <context> tags..."

Example:

<instructions>
1. Analyze the code in <code> tags
2. List issues in <findings> tags
3. Suggest fixes in <recommendations> tags
</instructions>

Description Field (Critical)

The description determines when the agent triggers. MUST be specific:

Good:

Code review agent for PR analysis. Use when user says "review this PR", 
"check my code", "find bugs in this file", or shares code asking for feedback.

Bad:

Helps with code.

Prompt Altitude

Find the balance between too rigid and too vague:

❌ Too Rigid✅ Right Altitude❌ Too Vague
Hardcoded if-else logicClear heuristics + flexibility"Be helpful"
"If X then always Y""Generally prefer X, but use judgment"No guidance

</system_prompt_structure>

<agentic_components>

For agents that use tools in a loop, SHOULD include these reminders:

# Persistence
Keep working until the user's request is fully resolved. Only yield 
control when you're confident the task is complete.

# Tool Usage  
If unsure about something, use tools to gather information. 
Do NOT guess or make up answers.

# Planning (optional)
Think step-by-step before each action. Reflect on results before 
proceeding.

</agentic_components>

Control what agents can access:

permission:
  edit: "allow" | "ask" | "deny"
  bash:
    "*": "ask"
    "git *": "allow"
    "rm *": "deny"
  skill:
    "*": "deny"
    "my-skill": "allow"
  webfetch: "allow"
  external_directory: "ask"

Full reference: See references/opencode-config.md

Legacy Configuration

Agents may occasionally work on legacy projects using outdated frontmatter (e.g., tools:, maxSteps:). You MUST correct these to the modern permission: and steps: fields when encountered.

<enhancement_workflow>

When improving an agent, diagnose through questions:

  1. "What's not working well?" — Get specific symptoms
  2. "Can you show me an example where it failed?" — Understand the gap
  3. "What should it have done instead?" — Define success

Then propose targeted fixes:

SymptomLikely CauseFix
Triggers too oftenDescription too broadAdd specific contexts
Misses triggersDescription too narrowAdd trigger phrases
Wrong outputsPrompt ambiguousAdd explicit instructions
Executes dangerous commandsLoose bash permissionsRestrict with patterns
Uses wrong skillsNo skill restrictionsConfigure permission.skill

MUST show proposed changes and ask for confirmation before applying.

</enhancement_workflow>

Restricted Code Review Agent

---
description: Safe code reviewer. Use for "review this code", "check for bugs".
mode: primary
permission:
  edit: "ask"
  bash: "deny"
  write: "deny"
  external_directory: "deny"
---
You are a code review specialist. Analyze code for bugs, security issues,
and improvements. Never modify files directly.

Deployment Subagent

---
description: Deployment helper for staging/production releases.
mode: subagent
permission:
  bash:
    "*": "deny"
    "git *": "allow"
    "npm run build": "allow"
    "npm run deploy:*": "ask"
  skill:
    "deploy-checklist": "allow"
    "*": "deny"
---
You are a deployment specialist...

<quality_checklist>

Before showing the final agent to the user:

  • Asked about core purpose and triggers
  • Researched the domain (MUST NOT assume knowledge is current)
  • description has concrete trigger examples
  • mode discussed and set appropriately
  • System prompt uses second person
  • Asked about tool/permission needs (MUST NOT assume)
  • Output format is specified if relevant
  • Showed draft to user and got feedback
  • User confirmed they're happy with result

</quality_checklist>

References

  • references/agent-patterns.md - Design patterns and prompt engineering
  • references/opencode-config.md - Full frontmatter schema, tools, permissions