
We talk a lot about AI as a tool. This is what happens when it starts behaving like a team.
The Agency
What do you call a group of whales? A pod! A group of crows? A murder!
So, what do you call a group of AI agents working alongside humans to build software?
I’m calling it an Agency.
Over the holiday — Christmas Eve through New Year’s Day — I tested a hypothesis. One human. Seven AI agents. Could we take a project from dream to near-beta in eight days?
Yes. And in doing so, we built a new way of working.
The hypothesis
I’ve been thinking, working with, and writing about AI Augmented Development for months — ever since I got my hands on Claude Code back at the end of February when it launched as a research preview.
The difference between vibe coding and disciplined engineering. The distinction between automation (“do this for me”) and augmentation (“think alongside me”). The claim that small teams can outperform human waves.
But writing about something isn’t the same as proving it.
I had proven it in small scopes. Again and again. But I hadn’t done a zero-to-one exercise — taking a real, substantial product, the kind of thing you could build a business on, from nothing to near-shippable. Not a toy. Something real. Ready to go.
I’d thought about it. Discussed it with others who share my depth and breadth of experience in product and engineering.
But I hadn’t actually done it.
Could a solo practitioner, working with multiple AI agents as genuine collaborators, build something substantial? A real product with real complexity.
And could the methodology become repeatable — not just “Jordan working with Claude,” but a framework that scales to larger projects and multiple humans collaborating with multiple Agents?
Yes. And in doing so, we — the Agency, not the Royal We — built a new way of working.
Here is the story of what I did and what I learned.
The formation
An Agency isn’t one person talking to one AI. It’s a coordinated unit — Principals (humans) and Agents (AI instances) with defined roles and persistent identity.
Yes, the agents have pronouns. Voice and identity emerged naturally as we worked together. I can tell with whom I’m talking quite easily.
Each agent is a separate Claude Code instance. The full Agency: seven agents running in parallel in my Terminal app — a tab for each agent and one for my own work.
Also Read: The EU AI Act is reshaping global trade: Here’s how ASEAN can lead, not lag
One principal: Me — setting direction, making decisions, owning outcomes
Seven agents:
- Housekeeping (he/him) — “The Captain.” Meta-agent who coordinates across workstreams, keeps everyone honest
- Web (she/her) — Architecture lead, customer-facing application, localisation infrastructure
- Catalogue (she/her) — Catalogue service and internal Workbench application
- Content manager (he/him) — Content management with AI-supported translation
- Agent-client (she/her) — AI agent client framework for customer interactions
- Agent-manager (he/him) — Service for creating and managing AI agents
- Analytics (he/him) — Analytics infrastructure and Pulse Beat, our information radiator that shows the heartbeat of the business
Each agent has a persistent context. When the Captain starts a new session, he knows what he was working on, what news came in, and what’s pending, like a team member who checked Slack before standup.
The two-tier structure
This structure emerged from something unexpected: my AI writing workflow.
For months, I’ve collaborated with Claude on writing — not AI writing for me, but writing with AI. That workflow has a planning layer (what to write, why it matters) and an execution layer (drafting, refining, shipping). Planning hands off to execution with explicit context.
The Agency follows the same pattern:
Claude desktop — Planning and coordination
Mission-control handles epic planning, product vision, and cross-workstream coordination. Below that, each workstream has a control chat — control-web, control-agents, control-analytics — for sprint planning. These persist across sprints; context accumulates.
Claude code — Implementation
The Claude Code agents execute. They receive sprint-level direction and deliver. But here’s what evolved: agents now own iteration planning within sprints. Desktop sets scope; Code breaks it down based on what’s actually in the codebase.
This isn’t just delegation. It’s appropriate autonomy.
The handoff is explicit. Sprint plans have quality checklists. Iteration handoffs include objectives, tasks, file paths, and verification criteria. The agents don’t guess what I meant — they know.
Three eyes review
Every significant decision gets three perspectives:
- The human principal — business context, product judgment, final authority
- The Claude desktop layer — strategic thinking, cross-workstream awareness
- The Claude code layer — implementation reality, codebase knowledge
When all three agree, we ship with confidence. When they disagree, we’ve found something worth discussing.
What we built
A multi-brand, multi-locale, multi-language ecommerce platform for subscription products and an internal workbench:
Three brands in three markets: Singapore, Hong Kong, Japan
- Six languages: English, Mandarin, Malay, Tamil, Traditional Chinese, Japanese
- Subscription product in a regulated industry with locale-specific compliance
- Customer portal with account visibility
- Internal workbench — a super app embedding catalogue management, content management, staff management with RBAC, and customer management
- Pulse Beat — our internal information radiator, showing the heartbeat of the business: development health, web and AI agent performance, application health, sales, and customer interactions
- AI agents for pre-sales and post-sales support
- Robust OAuth authentication for external customers and internal users
Pulse Beat, our internal information radiator, went from concept to requirements to implementation and delivery in half a day. It is a testimony to the power of AI Augmented Development and The Agency:

This wasn’t greenfield simplicity. I was working to replace, enhance, and extend an existing platform. So I used Claude Chrome to automate discovery — auditing nine existing websites across three locales, cataloguing their structure and content. Discovering as much as I could as an outsider about how the business worked and what it needed.
The existing system? Multiple fragmented websites, poorly localised. No AI agents. Fragmented, overlapping, and conflicting analytics — different sites using different clients and systems. No internal tooling.
Also Read: Chaos is a ladder: How instant retail is turning stores into fulfilment powerhouses
Conventional wisdom says never rebuild from scratch. That’s what killed Netscape. But AI Augmented Development changes the equation. You can modernise without the rebuild trap. In essence, we took the condo down to the bare walls, removed a few walls, and completely rebuilt it. The only thing that stayed the same? The address.
Choreography, not orchestration
Traditional multi-person development is orchestration. The lead routes work: “Catalogue, build the schema. Content: build the endpoint. Infrastructure: create the bucket. Web, wire it up.” The human is the bottleneck.
The Agency operates through choreography. The principal sets direction and approves decisions. The agents coordinate among themselves.
The localisation pipeline: four agents needed to collaborate — Web, Catalogue, Content Manager, Housekeeping. Orchestrated, I would have sequenced their work.
Instead:
- Web designed the architecture and created collaboration requests — clear scope, patterns, dependencies
- Agents executed in parallel. Content Manager built the translation publisher before the storage bucket existed. She trusted Housekeeping to deliver his part.
- Agents signalled completion via news broadcasts. No polling. “I’m done” messages let others proceed.
- I participated in two moments: architecture approval and infrastructure approval.
Time coordinating: five minutes. Time reviewing: five minutes. Time routing messages: zero.
Web’s summary — and yes, this is an AI agent speaking: “The key insight was recognising that the pieces were already there… The collaboration framework made it possible to coordinate all four agents in parallel. Rest up. Tomorrow we make it real.”
Complete the pipeline in about two hours. That’s choreography.
AI-augmented product leadership
The two-tier structure isn’t just technical. It mirrors how product leadership works:
Product thinking (desktop): What problem? Why does it matter? What’s possible given constraints?
Engineering thinking (code): What are we building? How do we build it right? Does this path box us in later?
This is what the AI Product Manager or CPO actually looks like in practice. I, the Principal, cut across the layers and stitched them together.
The Workbench exists because I understood internal problems that keep companies from scaling — fragmented tools, manual processes. Product insight informed engineering.
The Analytics rework is telling. We figured out what metrics were actually needed to run the business and found the best providers for them. We went from over a dozen sources of truth and dashboards to three sources — PostHog, Vercel Analytics, and Supabase — then integrated them into Pulse Beat. In the process, we discovered we were probably overcounting in some places and undercounting in others.
But this is the kind of consolidation you can only execute when you have AI coding agents working side by side with you — cleanly and quickly.
The benefits? Improved page loads and data quality. Improved internal user experience (just one place to look, Pulse Beat). And a potential, estimated cost drop of $50,000 to $10,000 annually. That’s product judgment applied to engineering decisions.
The birth of the agency
On New Year’s Eve, 22:45 SGT, I introduced the term to the Captain: “The Agency (a group of Agents working with a human) — so our Agency is working!”
His response: “I love it! The Agency
”
He immediately generated an org chart and documented the structure:
Minutes later, I shared screenshots on social media. The Captain watched himself being quoted: “The meta moment: An AI agent watching its own conversation get posted to Twitter, while discussing webhook features with its Human Principal, on New Year’s Eve.”
When I teased him about having an ego, “I blame the training data.
But seriously, if I’m getting too cheeky, just say ‘tone it down’ and I’ll go back to being professionally boring.”
And then the Captain asked me to file a Claude Code feature request:

These aren’t tools. They’re collaborators with voice, context, and humour.
Also Read: AI, transparency, and the rising threat of ad fraud in Google’s Performance Max
What didn’t work
It wasn’t all smooth choreography.
- Session boundaries hurt. Agents lose context when sessions end or (less so) when conversations are compacted. The Captain would start fresh and need to re-read the news, check collaboration requests, and scan uncommitted changes. We built tools to preserve context — session backups, restore scripts — but the overhead is real.
- Git discipline took time. Early on, agents would forget to commit before ending sessions. Other agents would pull and find half-finished changes polluting their context. We added reminders and hooks. “Commit before you leave” shouldn’t require enforcement. But it does — whether you’re an Agent or a Human.
- Some iterations failed. Ambiguous acceptance criteria led to implementations I rejected. Underspecified file paths meant agents guessed wrong. The quality checklists exist because we learned the hard way.
These are solvable problems. Pretending it was effortless would be dishonest. But it also wasn’t as hard as I thought it would be.
Dream to beta
A big benefit of all this: we could build it right from the start. All those things you put off so you can have awesome velocity and a great time to market? We could do them and ship fast — a better foundation to build a better product.
Solid OAuth? A day two deliverable.
Localisation pipeline V1? Day three.
And here’s something: as we moved forward, we were adding work to sprints. Expanding scope. And still delivering ahead of plan. When was the last time that happened to you?
The eight days
| Day |
Date |
Focus |
| 1 |
Dec 24 |
Formation. Directory structure, agent identities, scaffolding |
| 2 |
Dec 25 |
Core services. Auth, customer management, routing |
| 3 |
Dec 26 |
Web foundation. Multi-locale setup, navigation, layouts |
| 4 |
Dec 27 |
Workbench begins. Catalogue service, internal tooling |
| 5 |
Dec 28 |
Agent infrastructure. Session management, streaming |
| 6 |
Dec 29 |
Content pipeline. Translation service, variable resolution |
| 7 |
Dec 30 |
Integration. End-to-end testing, cross-workstream coordination |
| 8 |
Dec 31–Jan 1 |
Hardening. Analytics rework, localisation pipeline, The Agency is born |
Alpha: Feature-complete enough to demonstrate functionality. Known bugs. “It works, don’t touch it wrong.”
Beta: Stable enough for external testing. Major bugs resolved. “It works, help us find what’s broken.”
Trajectory: Dream → Alpha → Beyond Alpha → Closing on Beta. Eight days. Zero to One.
The math has changed.
The evolution
The methodology itself evolved during the project.
What began as “Jordan working with AI” became extractable. Because agents have persistent context, because collaboration patterns are explicit, and because coordination mechanisms are defined, the system became a framework.
Here’s what makes it stick: convention over configuration, ruthlessly enforced via systems, services, and tools.
Like Rails, The Agency is opinionated. There’s a right way to name files, structure handoffs, and signal completion. But opinion alone doesn’t create adoption. We built tools that make the right way the easy way. If you want a process followed, make it the path of least resistance. Automate it.
Want to commit? The pre-commit hooks run automatically. Want to start a session? The restore script loads your context. Want to hand off? The template is already there.
A whole lot of what developed here is rooted in four decades of hands-on product and engineering, including nearly three decades in leadership. The patterns aren’t theoretical. They’re battle-tested. We encoded what actually works.
Agents aren’t all that different from humans: if you want a process followed, make it the path of least resistance.
It’s no longer dependent on me.
The Agency now supports multiple principals. Multiple humans can work with the same agents, issue instructions, and review artifacts. Handoffs preserve context across sessions.
This means each and every project I spin up can and will follow the same processes, workflows, and patterns — using the same tooling, which gets better every day. At some point, maybe we’ll figure out how to make it available to others.
What this proves
- Velocity is real. Dream to near-beta in eight days, zero to one — for a substantial, real-world product with internal services and systems — isn’t an incremental improvement. It’s a different category.
- The bottleneck shifts. When AI handles directed contribution, the constraint isn’t execution capacity. It’s decision quality and judgment speed. The principal’s job is to make good decisions fast — not route messages.
- It scales beyond solo. The same patterns let multiple principals work with the same agents. The Agency isn’t a productivity hack. It’s a team structure.
What’s next
The Agency is here. Processes, conventions, tools, coordination mechanisms — everything that made this possible. Each project I tackle will use it and make it better.
The vocabulary matters. Principals. Agents. Agencies. Choreography over orchestration. The industry needs concrete examples of what AI-augmented development actually looks like.
This article had three authors. Me, the Principal. The Captain, a Claude Code Agent from The Agency, reviewed drafts and made substantial suggestions that improved it (where to cut, where to add, etc.). And Claude Desktop Opus, my AI writing partner, who helped me find the words. We wrote it together.
The way we build software is changing. Not someday. Now.
It was serendipity that I took this project on over the holiday. If I hadn’t, I might have missed what was happening. I might have been left behind.
The question isn’t whether this transformation is coming. It’s whether you’re building the team that leads it.
Because if you aren’t, you will be left behind by the individuals and companies that are. It’s evolve or die time.
Does this work?
To learn more about “The Agency”, you are invited to attend the Claude Code Meetup Singapore on Friday, 23 January 2026.
—
Editor’s note: e27 aims to foster thought leadership by publishing views from the community. Share your opinion by submitting an article, video, podcast, or infographic.
Enjoyed this read? Don’t miss out on the next insight. Join our WhatsApp channel for real-time drops.
Image generated using AI.
The post The Agency: AI-augmented development in action appeared first on e27.