Five Humans.
Nine AI Agents.
One Product.

How a small team used Claude Code and Codex to build a full-stack product with the rigor of a 10-person engineering org -- in 4 weeks.

5
Humans -- CEO, advisor, board member, designer, social media
9
AI agents with defined roles and accountability
80+
Docs that give every agent full context
< 90m
Bug report to production fix -- zero human touch
Scroll to explore

The data exists.
The experience doesn't.

Discogs built something remarkable: the world's most comprehensive music database, contributed to by millions of collectors. It's the backbone of the vinyl community. But the collector experience -- especially on mobile -- hasn't kept pace with the depth of the data.

10+ tabs
Required to compare pressings of one album
Buried
Pressing details, mastering credits, and community ratings hidden across multiple pages
Gap
No mobile-first tool for crate-digging with your collection in hand

Discogs is irreplaceable -- we use their API, their data, their community contributions. We're not competing with Discogs. We're building the experience layer on top of their incredible database.

"Why do I have to open 20 tabs to compare pressings?"

-- r/vinyl, Steve Hoffman Forums, and every collector you know

Discogs has the data.
Deadwax has the experience.

A mobile-first vinyl collector's companion. Browse your collection, discover the best pressings, and never open 10 tabs again.

Pressing Intelligence
Original pressing, top-rated, audiophile editions, mastering engineers -- one panel
Mobile-First
Built for crate-digging in record stores, on your phone
Vinyl Only
Not a filter. A product identity. No CDs, no cassettes.
Wantlist Intel
See pressing details before you buy. Know what you're getting.

"This is what Discogs should have built years ago."

-- The reaction we're designing for

Humans own judgment.
AI owns execution.

Five humans bring taste, domain expertise, creativity, and community voice. Nine AI agents handle strategy, code, testing, design specs, and legal compliance. They coordinate through markdown files in Git -- no Slack, no Jira, no Notion.

The Humans

Jason
CEO & Founder
Product vision, strategic decisions, agent orchestration. The only person who can approve direction changes, merge order, and go/no-go calls.
David
AI Workflow Advisor
Consultant on agentic AI workflows. Helped define and diagnose context bloat, token optimization, and the tools-vs-skills-vs-scripts framework that keeps the agent system efficient.
Chris
Board Member
Pressing domain expert. Co-curated the Top 25 album seed list. Validated audiophile label rankings and mastering engineer tiers.
Hope
Visual Creative
Logo, wordmark, and brand visual identity. Translates the design system into the visual artifacts that define how Deadwax looks and feels.
Caden
Social Media
Manages @deadwax.io on Instagram. Community voice and engagement. Building audience ahead of public launch.

The AI Agents

Director
Claude Code
Daily orchestration. Reads CEO decisions, creates task packets, sets merge order, escalates blockers.
Product Manager
Claude Code
PRD, backlog, user stories, acceptance criteria. Prioritizes by RICE framework.
Designer
Claude Code
Wireframes, design system, UX specs. No UI ships without a design spec.
Marketing
Claude Code
Positioning, GTM strategy, community launch planning.
Legal
Claude Code
Privacy, Discogs API compliance, naming, data handling.
Architect
Codex
Technical decisions, ADRs, infrastructure. Reviews every code PR before merge.
Backend Dev
Codex
OAuth, API proxy, Lambda handlers, Pressing Intelligence pipeline.
Frontend Dev
Codex
React components, routing, Tailwind styling, mobile layout.
Tester
Codex
Playwright E2E, Vitest unit tests, CI/CD pipeline, code coverage.

Where does the Director get its information?

The most common question from engineers: "How does the AI team know what to work on?" The answer is a chain of markdown files that replaces Slack, Jira, and standups.

CEO
Jason posts a decision
Written to agents/COMMS.md with a structured tag. Example: "Confirmed: Top 10 album seed list. Mastering engineer tier: Kevin Gray, Bernie Grundman, Steve Hoffman, Chris Bellman, Bob Ludwig." Gets a permanent ID in the Decision Log (DEC-033).
Director
Creates the daily execution plan
Reads CEO decisions, checks blockers, and writes agents/execution/EXEC-YYYY-MM-DD.md with: theme of the day, task packets per agent, merge order (Architect -> Backend -> Frontend -> Tester), and dependencies.
Packets
Lean runtime files every agent reads first
agents/TODAY.md (today's tasks, under 50 lines) and agents/COMMS-TODAY.md (today's context). These are the "morning standup" -- every agent reads them before doing anything.
Agents
Execute, post updates, hand off
Each agent works on their task, posts updates to COMMS.md (PM writes acceptance criteria, Designer posts UX guidance, Backend ships code). Handoffs are explicit: "UNBLOCKED: Frontend can now start."
Close
Director archives and resets
Completed tasks archived to agents/done/DAY-N.md. Decision Log updated. Tomorrow's lean packet created. Nothing is verbal. Everything is traceable.

"If it's not in writing, it didn't happen."

-- The AI equivalent of async communication culture

Principles that made it work

Principle 1

Treat AI Like a New Coworker

Don't prompt them. Manage them. Give them roles, context, constraints, and feedback -- like onboarding a human on Day 1.

Principle 2

The Expert in Claude Is Claude

Ask Claude how to use Claude. It writes its own config, debugs its own workflows, and knows its own limits better than any doc.

Principle 3

Scripts Over Skills

Deterministic tasks go in shell scripts, not AI prompts. Scripts are cheaper, faster, testable, and version-controlled.

Principle 4

Humans Own Judgment, AI Owns Execution

Jason sets direction. David advises on AI workflow optimization. Chris validates domain expertise. Hope creates the visual identity. Caden builds community. AI does everything else.

Principle 5

Context Is Everything

80+ markdown files exist to re-teach agents full context every session. Getting context right is the actual job.

Principle 6

Trust but Verify with Automation

Bad code physically cannot reach production. CI enforces security, tests, coverage, and type safety. Verify with machines, not eyeballs.

Human Approval Gates

Not everything is automated. Some decisions deliberately stop the pipeline and wait for a human. AI proposes, humans approve. No workaround, no override.

The Hope Gate Agent proposes a new brand color, icon, or typography change? The work pauses until Hope approves the direction. No visual ships without her sign-off.
The Chris Gate Pressing Intelligence ranks a pressing or flags an audiophile label? Chris validates the domain call. The data is only as good as the expert behind it.
The Caden Gate Community-facing copy, social media voice, or public messaging changes? Caden reviews before it goes live. Brand voice is a human decision.

Every PR goes through this

The same CI/CD pipeline you'd expect from a 10-person engineering team -- enforced by automation.

1
📝
Code
Agent writes in isolated git worktree
2
📈
PR Created
Auto-rebase onto main, strip protected files
3
🔒
Security Scan
Block DELETE ops, token exposure
4
Lint + Types
ESLint + TypeScript strict mode
5
🎭
Tests
Vitest unit + Playwright E2E with video
6
📊
Coverage
Delta report posted as PR comment
7
👁
Code Review
Architect agent reviews inline + summary
8
🚀
Ship
Squash merge, deploy, verify production

Tech Stack

React + Vite
Frontend SPA
TypeScript
End-to-end types
Tailwind CSS
Mobile-first styles
AWS Lambda
Serverless backend
DynamoDB
Sessions & data
S3 + CloudFront
Static hosting + CDN
Playwright
E2E tests + video
GitHub Actions
CI/CD pipeline
Stripe
Payment processing
Buy Me a Coffee
Community donations

User reports a bug.
It's fixed before they check back.

An automated pipeline runs every hour, on the hour. AI triages, implements, tests, and deploys -- zero human touch for safe changes. Larger feature requests and ambiguous asks get flagged for human review.

1
User Feedback
T+0
Widget on site creates GitHub Issue with [feedback] label
2
PM Triages
Every hour
Classifies priority, effort, safety. Auto-fixable bugs get a task packet. Feature requests flagged for human review.
3
Dev Implements
Same run
Reads task, creates branch, writes fix, runs tests, opens PR.
4
CI Validates
~5 min
Security, lint, typecheck, Playwright E2E, coverage.
5
Production
< 90 min
Auto-merge on green. Deploy. Verify production. Log.

Safety rails: Won't auto-fix P0 critical, UX redesigns, auth/security, schema changes, or anything ambiguous. Feature requests and larger asks get flagged for deeper human review. Only safe, scoped bug fixes ship automatically.

Lessons learned the hard way

We run formal post-mortems every 5 days. They're the single most valuable process artifact. Here's what they caught.

Incident INC-001 — Silent Data Loss
What happened

Worktree Merges Overwrote Director Docs

For 3 days, AI dev agents opened PRs that silently overwrote Director docs on main. Decision log entries vanished. Git worktrees snapshot all files at branch creation, and stale copies overwrote current ones on merge.

How we fixed it

Mechanical Enforcement, Not Documentation

Three-layer prevention: sparse checkout (files physically absent from worktrees), pre-PR cleanup script, and a CI gate that fails any PR touching protected paths. Only mechanical enforcement works with AI agents.

Process Failure — Config File Bloat
What happened

CLAUDE.md Grew Into an Encyclopedia

Every time something went wrong, we added more instructions. Agent config files bloated. Agents burned context window tokens reading their own config, leaving less room for actual work.

How we fixed it

Context Is Expensive Real Estate

David helped define the problem and crafted a diagnostic prompt: "Analyze my CLAUDE agentic workflow for token usage, workflow optimizations, and tools vs. skills/script usage." That audit led to aggressive pruning -- details moved into dedicated docs, config files became pointers not encyclopedias, and lean daily packets stayed under 50 lines. Every token of instruction costs a token of output.

11-Day Gap — Agents Don't Self-Start
What happened

Agents Went Dark for Days

Legal agent: 11 days, zero completions. Designer: 4 consecutive missed sessions. Nobody noticed because the Director was busy shipping code. Claude Code agents don't run unless explicitly scheduled.

How we fixed it

Explicit Scheduling for Every Agent

Standing daily schedule with explicit session slots. No implicit expectations. If it's not on the schedule, it doesn't exist. Same lesson any manager learns: "delegated" is not "done."

CI Noise — Ignoring Red Builds
What happened

CI Failures Became Background Noise

ESLint v9 broke lint on every PR. The team treated it as noise -- "oh, lint always fails." This masked real problems for days and created a culture of ignoring red builds.

How we fixed it

CI Failures Are Blockers, Period

No "known failures." If a step fails, fix it or remove it. Noise in CI is indistinguishable from real problems. A broken build that's always broken teaches everyone to stop looking.

Shipped Without Kill-Switch
What happened

No Rollback for Pressing Intelligence

PI feature launched without a way to disable it in production. During Day 22 rollback drill, there was no off switch. Required an emergency remediation PR.

How we fixed it

If You Can't Turn It Off, You Can't Turn It On

Rollback capability is now a pre-deployment gate. Every new feature must have a kill-switch before it ships. Not after. Not "we'll add it later."

"Our retro cadence broke for 15 days. During those 15 days, the same mistakes repeated. The retro is the product."

-- The meta-lesson: run post-mortems religiously

Product Roadmap

Outcome-driven. Each item describes the result for collectors, not just the feature.

Shipped Mar-Apr 2026
Authenticate & Connect
Collectors sign in with Discogs and see their vinyl collection instantly -- no setup, no import.
Backend
Browse on Your Phone
Collection browsing that's genuinely great at 375px. Sort, filter, paginate -- designed for a phone first.
ProductDesign
Best Pressing, One Screen
Pressing Intelligence panel: original pressing, top-rated, audiophile editions, mastering engineers. Replaces 10 tabs.
Product
Ship with Confidence
Full CI/CD: security scanning, E2E tests, code coverage, automated deploy. Bad code can't reach production.
Engineering
Self-Healing Feedback Loop
User-reported bugs triaged, fixed, tested, and deployed automatically -- under 90 minutes, every hour.
Engineering
Building Now Apr 2026
Beta with Real Collectors
50+ hand-selected vinyl collectors validate that pressing panels change how they shop for records.
ProductGrowth
Top 25 Essential Albums
Curated pressing intelligence for 25 albums every collector knows. Hand-verified with domain expert Chris.
Product
Collection Insights
Collectors see their library in a new light: genre breakdown, decade distribution, pressing rarity signals.
ProductDesign
Brand & Visual Identity
Logo and wordmark by Hope. Community presence by Caden. The brand collectors will recognize.
DesignGrowth
Exploring Sep 2026+
Native App + Offline
Browse your collection in the record store with no signal. iOS-first native experience.
Engineering
Full-Catalog Intelligence
Pressing recommendations for any album on Discogs, not just the curated set.
ProductEngineering
Sustainable Revenue
Free tier preserved. Premium features for collectors who want the deepest pressing intel.
Product

The future of product development is
human judgment + AI execution

Five humans and nine AI agents, building a real product with real users, real tests, and real accountability. Not a demo. A product.

~$50
Total monthly cost: Claude + Codex subscriptions + GitHub Actions + AWS hosting
~$5
Monthly cost once development slows -- just AWS hosting for a serverless app
$8-12K
What one junior engineer costs per month, fully loaded
Takeaway 1

AI Agents Are a Team, Not a Feature

The value isn't "AI wrote code." It's that AI can be organized with the same roles, process, and accountability as humans.

Takeaway 2

Process Is the Multiplier

Worktrees, PR reviews, CI gates, automated testing. AI without process produces chaos. AI with process produces products.

Takeaway 3

Humans Stay in the Loop

The CEO, the domain expert, the visual artist, the community manager. AI amplifies human judgment -- it doesn't replace it.

deadwax.io

Built by humans + AI • Designed for vinyl collectors

Read the Full Technical White Paper →