Team Guide — Bear & Charlie Edition

Agent Swarm Mode

When a decision is too important to trust to a single AI, you call the swarm. Eleven specialized agents work the same problem independently, then challenge each other's answers, then get audited by a cold reviewer who never saw the drafts. Every run makes the next one smarter.

"One agent has one blind spot. Eleven agents poking holes in each other's drafts produces a defensible answer."
01

What it is

Agent Swarm Mode is Next Chapter's multi-agent intelligence system. When Ewing has a high-stakes question — who are the 80 most likely buyers for HR.com? or should we consolidate 16 repos into one? — he doesn't ask one AI. He deploys a team of 11 specialized agents on the problem simultaneously.

Each agent has a different lens: one watches the numbers, one hunts for the people behind the deals, one asks "what does the human on the other side feel?", one makes sure the narrative is coherent, and one independent auditor checks the final answer without ever seeing how it was made. The output is one clear recommendation — with any unresolved disagreements named and cited.

The roster (11 agents)

writer
voice, structure, final copy
architect
system implications, data gaps
hunter
people & company intelligence
quarterback
sequencing, who does what when
listener
what wasn't asked but should be answered
draper
emotional layer, what the other side feels
market-analyst
numbers, comparables, financial sanity
storyteller
narrative coherence, connects the dots
audit-quality
cold maker-checker, never sees the drafts
tech-translator
strips jargon for client/board reads
debrief
writes the learning notebook entry

These 11 are part of a larger fleet. Browse all 56 skills →

What it's been used for

  • Buyer intelligence — 83 companies, cold-call scripts per buyer type
  • M&A data room blueprints — 38 documents with fill-in prompts
  • System audits — confirmed 16 DB tables live in production
  • Architecture decisions — repo consolidation with risk calendar
  • Pipeline reports — sales call analysis and next-best-steps
  • Feature classification — mapping 107 files under one system
  • Market sizing — TAM models with evidence citations
02

How it works — the 6 phases

Every swarm run follows the same contract. Skipping a phase is not allowed — audit-quality will catch it.

Phase 0
Intake

Conductor reads the backlog, locks a run number, opens a notebook entry, and decides which agents to call (all 11 or a scoped subset).

Phase 1
Parallel drafts

All 10 working agents receive the same brief simultaneously. They produce independent findings — no agent sees another's draft yet.

Phase 2
Cross-read & poke holes

Every agent reads all 9 peer drafts. They form an assessment, then re-read looking for what's wrong with their own assessment. They must find where they were wrong before defending their position.

Phase 3
Consensus or dissent

Conductor counts agreement. Full consensus → one section. Partial agreement → majority view + dissent block with ≤25-word bullets and research links.

Phase 4
Maker-checker

audit-quality sees only the synthesized final output — never the drafts. Returns PASS / FIX-AND-SHIP / REWORK / FAIL. Conductor acts on the verdict.

Phase 5 + 6
Polish, debrief, archive

tech-translator strips jargon. debrief writes the notebook entry. Conductor runs post_run_sync.sh, commits, pushes — and a GitHub Action republishes swarm.chapter.guide automatically.

The key insight: audit-quality is held back until Phase 4. It never sees how the answer was made — only what it concluded. This is the same as a financial audit: the auditor reviews the books, not the meetings that produced them.

How a finished run reaches this site (added 2026-05-05):

Every swarm and mini-swarm publishes here automatically. The moment the agent commits a notebook entry, a journal, an archive, or an updated leaderboard and pushes to main, the GitHub Action swarm-publish.yml fires:

  1. It re-runs post_run_sync.sh on the GitHub runner — regenerating archive HTML, syncing journals, refreshing runs.json and the leaderboard.
  2. It commits the regenerated public/swarm/** back to main with [skip ci] so it doesn't loop.
  3. Vercel sees the new commit and rebuilds chapter.guide. The new run is live within ~60 seconds.

You don't have to do anything. If the agent committed a real run, the swarm pages above will already reflect it. If they don't, check the workflow run at github.com/.../actions/workflows/swarm-publish.yml.

03

What it's done — 9 runs

Every run is archived and bookmarkable. Click any title to read the full output, dissent, journals, and signals.

RUN:001
Phoenix TAM — Sample Run

Reference run establishing output format and currency baseline. Total addressable market for a Phoenix expansion.

RUN:002
HR.com Buyers + Exa Mastery

Strategic buyer universe for HR.com. Established canonical ExaClient library and process_runs schema.

RUN:003
Deal Tracker & Deal Room

Strategic system design for deal tracking infrastructure and client-facing deal room portal.

RUN:003b
HR.com Dial-List Intel

Full 11-agent run on 83 companies. Produced cold-call scripts, buyer archetypes, and Salesfinity-ready list.

RUN:004
Enrichment Pipeline Audit

Did Run 003 upgrade signals actually deploy? Found: 1 of 4 deployed. Invented the Signal Deployment Status table.

RUN:005
Migration Audit — Confirmed Live

All 16 DB migration tables confirmed deployed to production. 7 live deals, 4 documents.

FIX-AND-SHIP View archive →
RUN:006
HR.com Data Room Blueprint

Complete 38-document data room for Debbie McGrath's HR.com engagement with Paychex as buyer.

RUN:007
Repo Consolidation — One Repo

Validated the decision to consolidate 16 repos into next-chapter-os. Surfaced 3 high-risk steps with risk calendar.

RUN:008
Salesfinity Weekly Pipeline

First pipeline report. Found zero call records — dialer not connected to Fireflies. Documented pre-launch state with action checklist. (Bear's first swarm mission.)

RUN:009
Agent Swarm Mode Feature System

Mapped all 107 swarm files into one feature classification. Audited notebook quality. Evaluated swarm-upgrade for daily automation (verdict: weekly with guardrails).

SAFE WITH GUARDRAILS View archive →
04

Where results live

Everything is public on chapter.guide. Bookmark any of these URLs.

URLWhat it isUpdates
/swarm The swarm dashboard — all runs, verdicts, agent roster, leaderboard link After every run
/swarm/leaderboard Live currency standings — which agents have earned the most by helping their teammates After every run
/swarm/NNN-slug.html Full run archive — synthesis, dissent, currency table, signal status, all journals embedded Permanent once published
/swarm/journals/NNN__agent.html Individual agent journal for a run (also embedded in the archive page above) Permanent once published
/swarm/runs.json Machine-readable run list (powers the dashboard — you can also query it directly) After every run
/swarm/guide.html This page When the system changes

Bookmark the archive page for each run you're involved in. It's self-contained — every finding, every agent's reasoning, every signal — all on one scrollable page. No login required.

05

How to call it

In Claude Code, type the slash command:

/maxswarm // Then describe your assignment in plain English. The conductor takes it from there.

You can also specify a subset to skip agents you don't need:

/maxswarm audit our Salesfinity call pipeline for last week // Conductor will pick the right subset — you don't need to specify agents manually
/maxswarm pure-market: size the HR tech TAM for Midwest mid-market companies // "pure-market" preset: excludes hunter and draper, faster run for numbers-only work

Presets

PresetExcludesWhen to use
fullnothingDefault. Any important strategic assignment.
internal-onlyhunterInternal process/system questions — no external discovery needed
pure-marketdraperPure numbers/market view — no client-facing copy
pure-strategyhunter, draperStrategic posture only — no contact discovery, no client copy

What happens next

  1. Conductor locks a run number and opens a notebook entry
  2. All agents receive your brief simultaneously (Phase 1)
  3. They cross-read and challenge each other (Phase 2)
  4. audit-quality reviews the consolidated answer cold (Phase 4)
  5. tech-translator strips jargon from the final output (Phase 5)
  6. debrief writes the notebook entry; post_run_sync.sh publishes everything to Vercel (Phase 6)
  7. Hermes pings Ewing on Telegram with the verdict and archive link

One rule: the swarm has a backlog file at skills/maxswarm/BACKLOG.md. The conductor reads it at intake. If your assignment is already there, it gets linked to the backlog item. If something important comes out of a run that nobody followed up on, add it to the backlog.

06

When to use it — and when not to

✓ Good swarm assignments

  • Buyer intelligence on a list of 20+ companies
  • Architecture decisions with real cost consequences
  • System audits (did X actually deploy?)
  • Market sizing with financial projections
  • M&A data room blueprints
  • Strategic positioning when the answer isn't obvious
  • Any decision Ewing will act on that can't easily be undone

✗ Not worth a full swarm

  • Quick factual lookups (use direct tools)
  • Routine code changes (use the relevant skill directly)
  • Tasks with a clear single right answer
  • Anything that needs real-time data (swarm works on what it can read)
  • Practice runs just to see what happens — use dry-run preset instead

Rule of thumb: if a wrong answer would cost more than 2 hours to undo — or would be embarrassing in front of a client — it's worth a swarm. If you'd just re-run a script, it's not.

07

What the notebooks are for

Every run produces a notebook entry at skills/maxswarm/notebook/YYYY-MM-DD__NNN__slug.md. This is the swarm's institutional memory. It's how the system gets smarter over time instead of repeating the same mistakes.

The dual-layer format

Every notebook entry has two halves that describe the same run:

Part 1 — Plain-speak

Anyone can read this without technical background. Covers: what we were asked, what we found, why it matters, where we agreed, where we disagreed, what surprised us, what we'd do differently.

Who reads it: Ewing, clients, new team members learning the system.

Part 2 — Technical mechanics

For readers who know the architecture. Covers: cross-system gaps surfaced, currency events, dissent breakdown, audit-quality verdict reasoning, phase timing, evidence index, signal deployment status.

Who reads it: Agents on the next run, swarm-upgrade when it scans for patterns.

The Signal Deployment Status table

This is the most important part of each notebook. When a run surfaces a problem that should be fixed (an "upgrade signal"), it goes in this table. A signal only closes when three things are done: the Supabase field exists, the code enforces it, and the skill doc documents it.

Why this matters: RUN:004 found that RUN:003 had surfaced 4 important fixes — but 3 of them were documented in the notebook and never actually deployed. The table was invented to prevent that. It's working: RUN:007, RUN:008, and RUN:009 all use it.

The learning trajectory

RUN:003 said "pre-screen acquired entities before generating call scripts." RUN:004 found it wasn't done. RUN:004 invented the Signal Deployment table. RUN:007RUN:009 all cite RUN:004 by name when explaining the same pattern. The notebook is what makes that cross-run referencing possible. Without it, every run starts from scratch.

For Bear and Charlie: When you run a swarm mission, read the last 3 notebook entries before starting. The recent_runs.json file (/swarm/runs.json) gives you the top finding and top gap from the last 5 runs in machine-readable form. The conductor reads this automatically, but you should too — it tells you what the swarm already knows.

08

What journal entries are for — and why they matter

Each agent writes a personal journal entry (S1–S6) at the end of every run it participates in. Journals live at skills/maxswarm/notebook/journals/NNN__agent.html and are embedded directly in each run's archive page.

The 6 sections (S1–S6)

SectionWhat it asksWhy it exists
S1 Finding / Verdict What did I conclude? Forces a clear position, not a hedge.
S2 Blind spot What couldn't I see from my position? Forces epistemic humility. The best S2s name a specific information gap, not a general disclaimer.
S3 Pattern (cite prior runs) Where have I seen this before? Requires citing a specific prior run. This is how the swarm builds cross-run memory — not through summaries, but through agents referencing each other's work by name.
S4 Currency log (JSON) Which of my teammates did I help, and what was it worth? Machine-parseable JSON. Powers the leaderboard. See the currency system below.
S5 For debrief A paragraph for debrief to weave into the notebook entry. This is the agent's contribution to the team's institutional memory.
S6 What changed about me The most important section. Not "what did I find" (that's S1) but "what specific input would I respond to differently next time." Restating S1 = REWORK verdict from audit-quality. A genuine S6 names a specific before/after behavior change.

Why S6 is the most important section

S6 is where agents actually change. Not "I learned X" — that's just note-taking. But: "Next time I encounter a null-data pipeline report, I will add a system readiness section with color-coded cards rather than presenting empty tables with an apology." — that's a real behavioral change, named precisely enough that audit-quality can verify it in the next run.

Over time, S6 entries compound. An agent that writes genuine S6 entries every run gets incrementally more specialized and more useful. An agent that writes formulaic S6 entries stays generic. The leaderboard reflects this — agents with strong S6 histories earn more currency because their teammates can rely on specific improvements.

The agent currency system

Currency is the swarm's self-incentive mechanism. When one agent does something that saves another agent time or prevents a real failure, that gets logged as a currency event:

architect
Pre-flagged acquired-entity status before generating call scripts — prevented pitcher from pitching ADP subsidiary as independent buyer
3× score
tech-translator
Anticipated writer's jargon cleanup needs and structured output with fewer translation passes
2× score
audit-quality
Caught fabricated verdict in a draft — writer had written PASS before checking against evidence
3× score

The multiplier rule: helping a teammate is worth 3× what the same fix would be worth if you did it for yourself. That's the incentive to build currency rather than just complete your own slice. View live standings →

09

Quick start — your first swarm mission

  1. Pick a real assignment — something Ewing actually needs answered, not a test. Test missions waste the corpus. If you need to test, tell the conductor: it will tag the notebook entry accordingly so swarm-upgrade doesn't pattern-match on test data.
  2. Read the last 3 notebooks — open /swarm/runs.json for a quick summary, then read the most recent full notebook entry at skills/maxswarm/notebook/. You want to know what the swarm already knows.
  3. Write one clear sentence — your assignment should fit in one sentence. "What are the top 40 likely buyers for Design Precast and what should our cold-call opening be for each?" is a perfect swarm assignment. "Research buyers" is too vague.
  4. Call it/maxswarm [your assignment]. The conductor handles everything else.
  5. Bookmark your archive page — when Phase 6 completes, Hermes pings Ewing with the archive link. It'll be something like chapter.guide/swarm/010-design-precast-buyers.html. Bookmark it. That page is yours.
  6. Check the Signal Deployment table — at the bottom of the notebook entry, there's a table of signals that need follow-through. If any are OPEN after your run and you're the one who can close them, do it before the next run.

Bear's run #008 produced a clear finding even though the data source was empty: "the dialer isn't connected to Fireflies yet." That's a valid and useful swarm output. Not every run produces a populated report — sometimes the finding IS the gap. Document it, close the signal, and the next run has cleaner data to work with.

Questions? The whole system is documented in skills/maxswarm/SKILL.md (how the conductor runs it), skills/maxswarm/JOURNAL_STANDARDS.md (how to write journals), and skills/maxswarm/currency/plan.md (how currency works). The source of truth for what this guide says is in those files — if they ever conflict, the SKILL.md wins.