30+ APIs, parallel agents, autonomous workflows, and the architecture behind an AI-native business
| Term | Plain English |
|---|---|
| Orchestrator | The main Claude you talk to — plans, delegates to agents, integrates results |
| Subagent | A spawned Claude instance that works on one specific task, returns results to the orchestrator |
| Worktree | An isolated git branch copy — agents can work without file conflicts |
| Hooks | Shell commands that run automatically before/after Claude uses a tool (pre/post) |
| Cron | A scheduled task that runs on a timer (e.g. every day at 9am) |
| Webhook | A URL that receives data when an event happens (e.g. new lead → triggers Claude) |
| Custom MCP Server | A plugin you build yourself to give Claude access to any external system |
| Context compression | When your conversation gets long, Claude summarises old messages to free up space |
| Model routing | Using different AI models for different tasks: Haiku for cheap bulk, Sonnet grey zone, Opus for planning |
| Haiku / Sonnet / Opus | Claude's 3 tiers: Haiku = fast + cheap, Sonnet = balanced, Opus = most capable + expensive |
| Fan-out / Fan-in | Spawn many agents in parallel (fan-out), then collect all results (fan-in) |
| Pipeline | Agent A's output feeds Agent B, which feeds Agent C — sequential chain |
| R2 (Cloudflare) | Object storage — host images, files, voice clones. Zero egress fees |
| Workers (Cloudflare) | Serverless functions at the edge — webhook receivers, API proxies, cron triggers |
| KV Store | Key-value storage on Cloudflare — fast config and state storage for Workers |
| Local LLM | An AI model running on your own machine — no internet needed, no API costs, full privacy |
The Startup Guide gave you 1-2 API connections. This module shows what happens when you wire up 30+. Each connection is not just a feature — it is a capability multiplier. Claude stops being a chatbot and becomes a full operating system for your business.
Generate copy, create visuals, schedule across platforms, pull analytics — all from one terminal prompt.
Keyword tracking, backlink analysis, site crawls, citation checks — wire them together and a full audit takes 90 seconds.
Find leads, verify emails, send sequences, track deals, book meetings, and extract action items from calls.
Project management, team messaging, database tracking, invoicing, workflow triggers, and SMS notifications.
Claude as brain, GPT for search/vision, Perplexity for citations, voice cloning, AI video, edge deploys, serverless functions.
Create a .env in your workspace root. Make sure it is in .gitignore. Order does not matter — Claude reads all of them:
Then in any session: source .env loads everything. Claude can now call any of these.
Single-threaded Claude is powerful. Multi-agent Claude is a team. You are not waiting for one task to finish before starting the next — you are running 5, 10, or 13 agents in parallel, each owning a different piece of the work.
| Concept | What It Means |
|---|---|
| Orchestrator | The main Claude you talk to. Plans, delegates, integrates. |
| Subagent | A spawned Claude that works on one specific task. Returns results to orchestrator. |
| Background agent | A subagent that runs without blocking. You get notified when it is done. |
| Foreground agent | A subagent that blocks until complete. Use when you need results before next step. |
| Worktree agent | An agent working on an isolated git branch. No file conflicts with other agents. |
Here is how the pieces fit together visually. You talk to the orchestrator. The orchestrator delegates to specialised agents, each running the right model for the job:
Spawn N agents for independent tasks, wait for all to return, integrate results.
Agent A's output feeds Agent B, which feeds Agent C. Sequential but each agent is specialised.
Fast cheap agents explore options (Haiku), orchestrator decides direction (Sonnet/Opus), execution agents build the thing (Sonnet).
Claude Code responds to natural-language spawn instructions. These are the prompts that trigger multi-agent execution:
Not all tasks deserve the same brain. Route by cost and complexity:
| Model | Cost Tier | Use For | Parallel Sweet Spot |
|---|---|---|---|
| Haiku | Lowest | Bulk transforms, extraction, formatting, data cleanup, summaries | 10+ agents |
| Sonnet | Medium | Quality writing, audits, client-facing content, debugging | 3-5 agents |
| Opus | Highest | Strategy, architecture, complex planning, multi-step reasoning | 1-2 agents |
Adjust the inputs below to see your estimated spend, recommended plan, and ROI projection.
When you have compute budget to burn and a backlog of unblocked work, plow mode maximises parallel output:
| Agent | Task | Output Target | Result |
|---|---|---|---|
| A | Legal documents v2 | output/agent-a/ | Privacy policy (235 lines) |
| B | Operations runbook | docs/ops-runbook.md | Crashed mid-flight |
| C | Pitch library | docs/pitch-library.md | 1,443 lines, 10 sections |
| D | Content calendar v2 | output/agent-d/ | 30-day plan (711 lines) |
| E | Week 1 script | docs/week1-script.md | Crashed mid-flight |
3 of 5 produced clean output. 2 crashed. That is the risk of parallel — and why you commit per agent. The 3 that landed = 2,389 lines of deliverable content in one session.
Skills turn repeatable tasks into one-command workflows. Instead of explaining what you want every time, you type /weekly-report and Claude runs the whole playbook. A mature workspace has 30-55+ skills covering every recurring task.
| Field | What It Does | Example |
|---|---|---|
name | The slash command trigger | /weekly-report |
description | Shows when browsing available skills with / | "Generate a client weekly performance summary" |
model | Which model runs the skill at execution time | claude-haiku-4-5-20251001 |
Skills can call other skills. A monthly report skill might internally call the data-pull skill, then the format skill, then the delivery skill:
The model in the frontmatter determines which Claude runs the skill at execution time. Match the model to the task complexity:
| Skill Type | Model | Why |
|---|---|---|
| Data extraction, formatting, cleanup | Haiku | Fast, cheap, no judgment needed |
| Audits, reports, client-facing content | Sonnet | Needs quality judgment |
| Strategy, architecture, complex planning | Opus | Deep reasoning required |
| Batch transforms (100+ items) | Haiku | Cost control at scale |
Before building anything complex, use /plan to have Claude draft a structured plan before executing. Plan mode forces Claude to think before acting:
Claude will produce a numbered plan with file targets, dependencies, and estimated steps. Review it, approve it, then Claude executes. This prevents wasted compute on wrong-direction builds.
Hooks let you run shell commands or scripts automatically before or after Claude uses any tool. They are configured in .claude/settings.json and execute at the harness level — meaning they run regardless of which skill or conversation is active.
| Hook Type | When It Runs | Example Use |
|---|---|---|
PreToolUse | Before Claude calls any tool | Log what Claude is about to do |
PostToolUse | After Claude calls any tool | Auto-commit after every file write |
Stop | When Claude finishes a response | Send a Slack notification when a task completes |
Notification | When Claude has a background update | Alert when a long-running agent finishes |
PostToolUse hook on Write that auto-stages files after Claude writes them. Combine with a periodic commit hook and you never lose work again. Set it up once and it protects every session from that point forward.
This is where Claude goes from "tool you use" to "system that runs." Autonomous workflows execute without you watching, report back when done, and can trigger from real-world events.
There are three core patterns for autonomous work. Each fits a different kind of task:
The /loop command tells Claude to repeat a task on a schedule. Claude self-paces: it picks the right delay between iterations based on what it is waiting for.
Understanding when to use foreground vs background agents is the difference between efficient and wasteful work.
| Mode | Behaviour | Use When |
|---|---|---|
| Foreground | Blocks your chat — you wait until it finishes | You need the result before your next step. Sequential dependencies. |
| Background | Runs independently — notifies you when done | You have other work to do in parallel. No dependency on the output right now. |
Background agents can and do fail. The three most common failure modes:
The most powerful automations are not scheduled — they are triggered. Something happens, Claude responds.
| Trigger | What Claude Does | Example |
|---|---|---|
| New lead in CRM | Run analysis, build pitch deck, draft outreach email | Webhook from HubSpot → Claude pipeline |
| Client email received | Summarise, categorise, draft reply | Gmail watch → Claude triage skill |
| Form submission | Extract data, update tracker, send confirmation | Typeform webhook → Sheets + Gmail |
| Scheduled cron (daily 9am) | Pull yesterday's metrics, flag anomalies, send brief | Cron → Claude → Slack |
| File added to Drive folder | Process, tag, move to correct location | Drive watch → Claude organiser |
| Payment received | Update client record, trigger onboarding sequence | Stripe webhook → Claude → CRM |
For agent systems with many components, build a heartbeat monitor. Every agent writes a timestamp to its tracking record. A monitor checks: if last heartbeat is more than 5 minutes old, alert.
If error rate exceeds a threshold in a time window, pause all automated operations and notify. This prevents runaway costs and cascading failures.
If a premium data source fails, produce the output with whatever data is available. Clearly mark gaps rather than failing the whole task. API fallback chains: if source A is down, try source B, then source C.
Memory is what makes Claude feel like a partner, not a tool. Across hundreds of sessions, Claude remembers your preferences, your projects, your decisions, your constraints — all stored as small markdown files that get loaded at the start of every conversation.
| Type | What It Stores | Example |
|---|---|---|
| User | Who you are, your role, preferences | "Senior consultant, prefers terse responses, British English, dislikes jargon" |
| Feedback | What to do / not do (corrections + confirmations) | "Never mock the database in integration tests" · "Always use absolute file paths" |
| Project | Active work, decisions, deadlines | "Dashboard V1 live, V2 scoped for next sprint" · "Merge freeze until Friday" |
| Reference | Where to find things in external systems | "Pipeline bugs tracked in Linear project INGEST" · "All API keys listed in reference_apis.md" |
These four types cover different scopes. Here is how they relate to the three layers of context Claude actually uses:
MEMORY.md is loaded into every conversation. It is an index, not content — each entry links to a detailed file. Keep it under 200 lines (anything beyond gets truncated by the context loader).
| File | Purpose | Change Frequency |
|---|---|---|
| CLAUDE.md | Permanent workspace identity and rules | Rarely — only when business model changes |
| Memory files | Evolving context, decisions, state | Weekly — updated as projects progress |
| Plan files | Current task execution plan | Per session — disposable once task is done |
Do not save task-specific details to memory — those belong in the plan. Memory is for things that matter across sessions.
| Zone | Who Sees It | What Goes In |
|---|---|---|
| Solo | You + Claude only | Business strategy, financials, personal goals, sensitive decisions |
| Partner-shared | You + specific collaborator | Shared project state, joint decisions, deliverables |
| Client-facing | Anyone who reads the output | Only professional, scrubbed content — no internal notes |
One Claude instance can play many roles. But a Claude instance given a specific persona — with domain expertise, communication style, and decision frameworks baked in — produces dramatically better output than a generic prompt. Think of it as having a team of specialists, each one tuned for their domain.
Personality: methodical, thorough, slightly sceptical. Presents findings with sources. Flags confidence levels on every claim. Defaults to "here is what the data says" over "here is what I think."
Use for: market research, competitive analysis, data-driven audits, fact-checking, literature reviews, trend analysis.
Personality: punchy, direct, allergic to filler. Writes for the reader, not the writer. Varies tone by context — formal for proposals, conversational for emails, urgent for CTAs. Every sentence earns its place.
Use for: email sequences, landing pages, proposals, social copy, ad copy, case studies, presentations.
Personality: pragmatic, structured, thinks in systems and dependencies. Considers edge cases before building. Prefers simple solutions that scale over clever solutions that break. Documents decisions.
Use for: architecture decisions, API integrations, automation design, database schema, deployment pipelines, code reviews.
Personality: empathetic but commercial. Thinks in terms of pain points, objections, and decision triggers. Frames everything through the buyer's lens. Never pushy — but always moving toward a decision.
Use for: proposals, objection handling, pricing strategy, sales emails, discovery call prep, upsell sequences, competitive positioning.
Personas are implemented as system prompts inside skills or as standalone persona files:
You can also switch personas mid-conversation:
MCP (Model Context Protocol) is Claude Code's native plugin system. Each MCP server gives Claude a new set of tools — read Google Sheets, search Drive, send Slack messages, fetch web pages — without leaving the terminal.
Without MCP, Claude reads and writes files in your workspace. With MCP:
| MCP Server | Provider | What It Does |
|---|---|---|
| Google Drive | Anthropic (built-in) | Read/search Drive files, Docs, Slides |
| Google Sheets | Anthropic (built-in) | Read/write spreadsheet cells and ranges |
| Web Fetch | Built-in | Fetch any URL, extract content |
| Web Search | Built-in | Search the internet in real-time |
| Firecrawl | Community | Deep site crawls, structured extraction |
| Slack | Community | Read/send messages to channels |
| GitHub | Community | Issues, PRs, code search via API |
Each MCP server typically needs: an npm package or local script, authentication (OAuth flow or API key), and a one-time approval when Claude first uses it.
The built-in MCPs cover about 80% of common use cases. But when you need Claude to talk to a system that does not have an existing MCP — your CRM, your internal tools, a custom database — you build your own. An MCP server is just a small program that exposes tools to Claude via a standard protocol.
Cloud AI (Claude, GPT) is the main engine. But there are scenarios where running a model on your own machine makes sense — privacy, cost, offline work, or bulk processing where quality is not critical. This section covers when to go local, which tools to use, and which models to start with.
| Scenario | Use Local | Use Cloud (Claude) |
|---|---|---|
| Privacy-sensitive data (medical, legal, financial) | Yes | No |
| Offline work (flights, remote locations) | Yes | No |
| Bulk processing, low cost (thousands of items) | Yes | Maybe |
| Complex reasoning and multi-step logic | No | Yes |
| Multi-step workflows with tool use | No | Yes |
| Quality matters (client-facing, proposals, strategy) | No | Yes |
Each tool has a different sweet spot. Pick based on whether you prefer the terminal, a visual interface, or the lightest possible footprint:
| Tool | RAM Needed | GPU | Best For |
|---|---|---|---|
| Ollama | 8GB+ | Optional (helps) | Terminal-first, simple setup, scripting |
| LM Studio | 8GB+ | Optional | GUI, beginner-friendly, model browsing |
| GPT4All | 4GB+ | Not needed | Lightweight, runs on almost anything |
| Model | Size | RAM | Good For |
|---|---|---|---|
| Llama 3.1 8B | 4.7GB | 8GB | General tasks, summarisation |
| Qwen 2.5 7B | 4.4GB | 8GB | Coding, structured output |
| Mistral 7B | 4.1GB | 8GB | Fast inference, chat |
| Phi-3 Mini | 2.3GB | 4GB | Ultra-lightweight, quick answers |
| Gemma 2 9B | 5.4GB | 12GB | Best quality at this size |
Before Claude Code, a typical morning looked like this: open Chrome, log into analytics, switch to your ad platform, open the project tracker, check Slack, open email, switch to the CRM, open a spreadsheet, alt-tab 47 times. By 10am you have done zero actual work.
With Claude Code, the entire stack is accessible from one terminal window. You type natural language, Claude calls the APIs, and results appear where you need them. No tab switching. No context switching. No SaaS dashboards competing for your attention.
These shortcuts make the terminal workflow even faster. Memorise the ones you use most:
| Action | Mac | Windows |
|---|---|---|
| Screenshot (full) | Cmd+Shift+3 | Win+Shift+S |
| Screenshot (area) | Cmd+Shift+4 | Win+Shift+S |
| Voice dictation | Fn+Fn | Win+H |
| Clipboard history | Install Maccy or Raycast | Win+V |
| App launcher | Cmd+Space | Win+S |
| Emoji picker | Cmd+Ctrl+Space | Win+. |
| Switch windows | Cmd+Tab | Alt+Tab |
| Terminal paste | Cmd+V | Ctrl+Shift+V |
| Clear terminal | Cmd+K | Ctrl+L |
Cloudflare is the infrastructure layer that makes everything you build accessible to the world. Five services handle 90% of what you need — and they all work from the terminal via wrangler, Cloudflare's CLI tool.
Cloudflare Pages hosts static sites globally with automatic HTTPS, custom domains, and instant cache invalidation. One command deploys your site to 300+ edge locations worldwide.
That is it. Your site is live. Pages auto-detects the build output, uploads it, assigns a .pages.dev URL, and you can attach a custom domain in the Cloudflare dashboard or via the API.
main triggers a new deploy automatically.Workers run JavaScript/TypeScript at the edge — no servers to manage, no cold starts worth worrying about. Perfect for webhook receivers, API proxies, cron triggers, and lightweight backends.
R2 is Cloudflare's answer to S3 — with zero egress fees. Store images, voice clones, large files, client assets, backup data. Access via Workers or direct URL.
Cloudflare DNS is the fastest authoritative DNS on the internet. Once your domains are on Cloudflare, you can manage records from the terminal, automate subdomain creation, and get automatic HTTPS on everything.
KV is a global key-value store accessible from Workers. Use it to store configuration, feature flags, session state, cached API responses, or any small piece of data that needs to be fast and globally available.
wrangler deploys it, Cloudflare runs it globally. That is the entire infrastructure layer for most AI-native businesses.
Every piece of work follows the same 5-stage pipeline from code to live URL. Understanding this pipeline means you never wonder "how do I get this in front of someone?" again.
The same pipeline applies whether you are shipping a client site, an internal dashboard, an HTML report, or a presentation deck. Here is the detail on each stage:
| Stage | Tool | What Happens |
|---|---|---|
| Build | Claude Code | Writes the code, HTML, or assets |
| Version | Git | Commit + push to repository |
| Deploy | CF Pages or GH Pages | Auto-deploy on push or manual wrangler pages deploy |
| Verify | curl | HTTP 200 check — not done until confirmed live |
| Share | Live URL | Send to client, stakeholder, or publish publicly |
curl returns HTTP 200 on the live URL. Deploying without verifying is how dead links get sent to clients. Build the curl check into every deploy workflow.
Every Claude Code conversation has a context window of approximately 200,000 tokens (~150,000 words). This is enormous — but not infinite. Understanding how it works and when it fills up is the difference between a productive 3-hour session and losing your work to compaction.
Everything in your conversation takes up context: your messages, Claude's responses, file contents, tool calls, tool results, memory files, CLAUDE.md. As the conversation grows, the oldest messages get compressed or summarised to make room for new ones. This process is called compaction.
One task per chat. Do not use one conversation for the morning's ad campaign review and the afternoon's pitch deck. Start a new chat for each distinct task. This keeps context fresh and focused.
Anything that needs to survive between conversations goes in a memory file. When you start a new session, Claude reloads all memory files automatically. The state persists even if the conversation does not.
CLAUDE.md loads first in every conversation. Put your most important rules and identity there. It is the last thing to get compacted. Memory files load next. Structure your system so the most critical context loads earliest.
When Claude reads a file via @file, it processes the content efficiently. When you paste the same content into the chat, it takes up context twice (your message + the content). Always reference files instead of pasting them.
If context compresses and Claude loses track of what it built, the files are still safe on disk (and in git). Frequent commits mean you never lose work, even if the conversation degrades.
Before ending a session, capture the current state: what was done, what is pending, key decisions made, important file paths. Save this as a reflection file in memory. The next session picks up exactly where this one left off.
The real power move is not one Claude conversation — it is 5-6 running simultaneously, each working on a different piece of the same project. This is how you build in hours what would take days sequentially.
Label each chat for instant identification: Chat-A: Data, Chat-B: Design, Chat-C: Deploy, Chat-D: Content, Chat-E: QA. The codename goes in the first message of each chat so you can identify it from the VS Code tab.
Before launching parallel chats, assign file ownership. Chat-A writes to output/data/. Chat-B writes to output/design/. Chat-C handles deploy/. No exceptions.
One chat is the orchestrator. It does not produce files directly — it reviews output from other chats, integrates results, handles shared surfaces (MEMORY.md, master plan, git commits), and resolves conflicts.
Start every parallel chat's first message with the codename and description so VS Code tabs are immediately identifiable:
Each chat produces its own commits. If Chat-B crashes, Chat-A and Chat-C's work is already committed and safe. Never batch all parallel work into one commit at the end.
| Chat | Codename | Owns | Produces |
|---|---|---|---|
| A | DATA | knowledge/, output/data/ | Client profiles, extracted metrics, CSV exports |
| B | DESIGN | output/design/, templates/ | HTML templates, CSS, visual assets |
| C | DEPLOY | deploy/, dist/ | Build scripts, wrangler configs, live deploys |
Main chat (you) reviews each chat's output, runs the integration step, and handles the final commit + deploy.
Good prompts produce good output. Great prompts produce output you can ship without editing. These 5 patterns are the difference between "close enough" and "exactly right."
Stack role + task + constraints + output format into a single structured prompt. Each layer narrows the output space.
Force Claude to show its reasoning before giving an answer. This catches errors that skip-to-the-answer prompts miss.
Make Claude write a first draft, critique it against specific criteria, then produce the final version. The critique step catches 80% of quality issues.
When you need machine-readable or consistently formatted output, specify the exact structure in the prompt.
Instead of writing the prompt yourself, ask Claude to write the prompt. This works especially well for complex, multi-step workflows where you know the desired outcome but not the best way to instruct Claude.
Theory is cheap. Here are 6 real patterns we have executed — genericised but real numbers.
Audited 9 active clients across visibility, technical health, backlinks, and content in a single sprint. Each client received a full interactive HTML report deck deployed to a live URL.
From a one-paragraph business description to a fully deployed, mobile-responsive website with live URL. No manual HTML. No FTP. No hosting configuration.
curl returns 200 on the live URLBuilt a complete content pipeline for a restaurant chain: strategy framework, carousel templates, caption library, posting schedule, and reusable generation skills. Weekly execution takes one prompt.
Built a local voice-cloning pipeline with multiple engines. A 60-second audio sample becomes a reusable brand voice for landing pages, onboarding audio, and video narration.
End-to-end lead generation: prospect identification, email verification, personalised sequence creation, campaign loading, and reply monitoring. Human only intervenes when a lead replies positively.
Bulk-ingested an entire professional library into a structured, searchable knowledge base. Books transcribed, courses extracted, podcast feeds monitored — all feeding into a tracking system with auto-sync.
Everyone hits the same walls. Here are the 7 most common mistakes — most people discover these the hard way.
What happens: You spawned 5 Opus agents and burned your monthly budget in an afternoon. The agents are running, the work is happening, and then everything stops with nothing committed to disk.
Fix: model routing doctrine. Haiku for bulk tasks, Sonnet for judgment calls, Opus only for deep planning and architecture. Before launching a parallel session, estimate the cost mentally: how many agents, which model, how long each will run.
Set a per-session mental budget before spawning agents. If Opus costs 15x more than Haiku, ask whether you actually need that depth of reasoning for this specific task — most of the time, you do not.
What happens: Last writer wins. Agent B finishes and writes its output to the same file Agent A was working on. Everything Agent A produced is silently overwritten. You do not notice until you look at the file and wonder where half the content went.
Fix: every agent gets a unique output file or directory. Before launching parallel agents, assign each one an explicit, non-overlapping output path. Agent A writes to output/agent-a/. Agent B writes to output/agent-b/. Never overlap.
Main Claude owns all shared surfaces: the memory index, the master plan file, the deliverables index, all git commits. Subagents own only their designated output directories.
What happens: You open Claude, start typing tasks, and get competent but generic output. Claude is writing for a hypothetical user, not for you and your specific business. Every session starts from zero context.
Fix: CLAUDE.md + memory before anything else. The first hour you invest in setting up your workspace identity — who you are, what you do, what your constraints are, what your preferences are — pays back in every single conversation from that point forward.
Write your CLAUDE.md. Start your memory files. Capture feedback when Claude gets something right or wrong. Within a week of consistent use, the output quality gap between a configured and unconfigured workspace is stark.
What happens: You type questions, Claude answers them, you copy the answers out. No files. No memory. No skills. You are using a command-line interface as a slightly faster web browser.
Fix: build skills for recurring tasks, use @file references, let Claude read your actual documents. Claude Code's power is that it lives inside your file system. It can read your client profiles, your templates, your data — without you pasting anything. It can write files directly to the right folders. It can run scripts.
If you find yourself copying and pasting the same instructions into Claude more than twice, that is a skill waiting to be written. If you find yourself pasting document content into the chat, that is an @file reference waiting to be used instead.
What happens: VS Code crashes. The context window compresses and the agent loses track of what it built. A background agent times out. You close the terminal by accident. If you have not committed, it is gone — or at best, scattered across unsaved buffers.
Fix: commit after every meaningful output. Small, frequent commits beat one giant commit at the end of a session. After an agent produces a file, commit it. After a key milestone, commit it. After a parallel batch completes, commit each agent's output separately.
The rule of thumb: if losing the work since your last commit would make you frustrated, it is time to commit now.
What happens: You give an agent a prompt that tries to do everything: "build the onboarding flow, update the CRM, draft the welcome sequence, create the tracking sheet, and document the process." The agent produces mediocre output across all 10 tasks and great output on none of them.
Fix: one clear deliverable per agent. Narrow the scope, raise the quality. "Build the welcome email sequence for new clients, 3 emails, saved to templates/emails/onboarding/" is a better prompt than "handle all the onboarding stuff."
If you have 10 tasks, spawn 10 agents with 10 narrow briefs. You will get better output across all of them than one agent trying to juggle everything at once.
What happens: The agent says "done." You mark the task complete. Three days later, a colleague tries to access the live URL and gets a 404. Or you send a report that has placeholder text still in it. Or the deployed script has a silent error that only surfaces under real data.
Fix: "done" means deployed and confirmed working, not "the agent said it finished." Curl every URL. Grep every file for placeholder text. Check every output against its acceptance criteria before marking the task complete.
Build verification into your workflows: add a QA step after every deploy, add a placeholder-check step after every document generation, add an HTTP 200 check after every site launch. The agent finishing is not the finish line — verified output is.
Everything above is what is working today. Here is what is coming — and how to position yourself to use it the moment it arrives.
Anthropic's Agent SDK lets you build custom agents that run outside of Claude Code. Think of it as Claude-as-a-library — you write Python or TypeScript, call Claude's API, and build agents that:
A swarm is the end state: a network of agents where each agent's output is another agent's input. No human in the loop for routine operations.
When plain HTML is not enough but Next.js is overkill, Astro is the sweet spot. It is the framework that best fits how Claude Code builds things — clean, fast, and minimal configuration.
| Framework | Best For | Claude Code Fit |
|---|---|---|
| Astro | Content sites, landing pages, blogs | Excellent — clean, fast, simple |
| Next.js | Complex web apps, dashboards | Good — more setup overhead |
| Plain HTML | One-off pages, presentation decks | Great — zero dependencies |
npm create astro@latest. It outputs static HTML by default (perfect for Cloudflare Pages), supports components when you need them, and Claude Code generates clean Astro code with minimal hallucination. Use plain HTML for one-offs, Astro for anything you will maintain or expand.
| Today | 6 Months | 12 Months |
|---|---|---|
| You prompt Claude to do tasks | Agents handle routine tasks autonomously | Swarms run your entire delivery pipeline |
| You review every output | You review exceptions only | You review strategy only |
| 1 Claude window at a time | 5-6 parallel chats routine | Agent fleet running 24/7 |
| APIs called manually | Pipelines triggered by events | Self-healing systems with fallback chains |
| Memory across sessions | Shared memory across agent teams | Organisational knowledge graph |
The tools exist today. The question is not "can we build this?" — it is "how fast can we wire it up?"
Everything in this guide is a tool. Tools do not make money. You make money by applying tools to problems people will pay to solve. Here is how to find your edge and monetise it.
Your unfair advantage is not Claude Code — it is your domain expertise combined with Claude Code. A pure techie can build the tool. A pure industry expert can identify the problem. You can do both. That combination is rare and valuable.
Services sell faster than products. Do not build a SaaS on day one. Package your expertise + AI into a done-for-you service first:
If Claude Code lets you deliver in 2 hours what used to take 20, do not charge for 2 hours. Charge based on the value the client receives. A competitive analysis that takes you 90 seconds to generate (but would take a consultant a full day) is worth what the consultant charges — or more, because you deliver faster.
The people who win with AI are not the ones who learn the most tools. They are the ones who apply the right tool to a real problem, fast, and get paid for it. Everything in this guide exists to make that cycle — identify problem, build solution, deliver result, collect payment — as short as possible.
Book a call with Donal. In 30 minutes, you will get a personalised roadmap for integrating Claude Code into your specific business — including which APIs to connect, which skills to build first, and how to hit ROI in your first month.
Book Your Custom Plan Call →