← All posts

AI-First Companies: Restructuring Around Harness Engineering

99% AI-written code, 3-8 deploys a day, two engineering roles. A field report from a company that actually rebuilt the loop.

April 6, 2026 9 min read Strategy
TL;DR

Adding AI to your existing workflow gets you 10-20% efficiency. Rebuilding the workflow around AI — harness engineering — is multiplicative. That means a monorepo agents can inspect, deterministic CI/CD, AI code review as a gate, and a self-healing triage loop. Two roles emerge: architects who design the harness, and operators who work the tickets it generates.

A team at CREAO ships a feature at 10 AM, A/B tests it by noon, kills it by 3 PM based on the data, and ships an improved version by 5 PM. Three months ago that loop would have taken six weeks. They didn't get there by dropping Copilot into the IDE. They took the engineering process apart and rebuilt it around agents.

OpenAI later named the pattern harness engineering: the team's job is no longer to write code. It's to give agents the ability to do useful work — and to make every failure a missing capability, not a missing effort.

AI-assisted is not AI-first

Most "AI-first" companies aren't. The engineer opens Cursor. The PM drafts specs in ChatGPT. QA tries AI test generation. The process stays the same and you get a 10-20% lift.

That's AI-assisted. AI-first means you redesign the process, architecture, and org on the assumption that AI is the primary builder. You stop asking "how can AI help engineers?" and start asking "how do we rebuild everything so AI builds and engineers provide direction and judgment?"

Vibe coding — prompt until something works, commit, repeat — produces prototypes. Production needs stability, reliability, and security even when AI writes the code. You build the system that guarantees those properties. Prompts become consumable.

Three bottlenecks that force the rebuild

Product management

PMs used to spend weeks researching, designing, and spec'ing. Agents implement features in two hours. A week-long planning cycle becomes the constraint. PMs either evolve into product architects working at iteration speed or step out of the build cycle.

QA

Same dynamic. Build time: two hours. Manual test time: three days. You've moved the bottleneck three meters downstream. Replace manual QA with AI-built testing platforms that test AI-written code.

Headcount

Competitors have 100× the people. You can't hire to parity. You have to design to parity — and that means all three systems (design, implement, test) need AI inside them. Leave one manual and it caps the entire pipeline.

The architectural move: one monorepo Claude can see

CREAO's old architecture sprawled across several independent repositories. One change meant edits in three or four places. From a human engineer's view: tolerable. From an agent's view: opaque. No cross-service reasoning, no local integration tests, no full picture.

They consolidated into a single monorepo for one reason: so AI can see everything. That's harness engineering in practice. The more of your system lives in a form an agent can inspect, validate, and modify, the more leverage you get. A fragmented codebase is invisible to agents. A unified one is readable.

Harness first: before you scale agents, build the harness. Fast AI without fast validation is fast-growing tech debt. The monorepo, the CI gates, and the observability come first. Everything else rides on top.

The deployment stack

Every code change runs a six-phase pipeline with no manual overrides: Verify CI → Build and Deploy Dev → Test Dev → Deploy Prod → Test Prod → Release. The CI gate checks types, lint, unit and integration tests, Docker builds, Playwright E2E, and environment parity. Determinism lets agents predict outcomes and reason about failures.

The key components:

The self-healing loop

This is the piece that makes the rest pay off.

Every morning at 9:00 UTC an automated health check runs. Claude queries CloudWatch, analyzes error patterns across services, and posts a summary in Teams. An hour later the triage engine clusters production errors from CloudWatch and Sentry, scores each on nine severity parameters, and creates Linear tickets with sample logs, affected users, endpoints, and suggested investigation paths.

The system deduplicates. Open ticket for the same pattern? Update it. Previously closed ticket recurring? Reopen as regression. When an engineer pushes a fix, the same pipeline processes it. After deploy, the triage engine rechecks CloudWatch. Errors resolved? Ticket auto-closes.

Every tool owns one phase. No tool tries to do everything. The daily cycle creates a loop where errors are detected, prioritized, fixed, and verified with minimal manual intervention.

Results over 14 days

People assume speed comes at the cost of quality. The opposite happened. Tighter feedback loops mean you learn more shipping daily than shipping monthly.

The two-role engineering org

Architect

One or two people. They design the SOPs that teach AI how to work, build the test and triage infrastructure, define system boundaries, and define what "good" looks like to agents. The role requires deep critical thinking — you critique AI, you don't follow it. When an agent proposes a plan, the architect looks for holes: what failure modes did it miss, what safety boundaries did it break, what tech debt is it accumulating.

Hardest role to hire for.

Operator

Everyone else. AI assigns tasks to people. The triage system finds a bug, creates a ticket, produces diagnostics, routes it. The human investigates, validates, approves the fix. AI writes the PR. The human checks for strategic risk, not line-by-line correctness.

The work still requires skill and attention. It no longer requires the architectural thinking the old model required.

Juniors adapted faster than seniors. Less to unlearn, tools that amplified their impact. Seniors had to watch two months of their old work get done in an hour. In this transition, adaptability matters more than accumulated skill.

The human side of the transition

Management collapsed. The CTO went from 60% of time on people to under 10%. He moved from managing to building — writing code, designing SOPs, maintaining the harness.

Most team conversations used to be alignment meetings, trade-off debates, priority arguments. Those conversations are necessary in the old model and exhausting. The team still talks, just about different things.

The anxiety is real. When the CTO stops talking to you every day, you wonder what your value is. Some people spend more time debating whether AI can do their job than doing the job. There's no clean answer — but the same principle applies to humans and AI: if AI makes a mistake, you don't fire the agent. You build better validation, clearer constraints, stronger observability.

AI-first doesn't stop at engineering

Ship features in hours while marketing needs a week to announce and you've moved the bottleneck. Same for a monthly product planning cycle. CREAO extended AI-native operations to every function: release notes from changelogs, feature videos via AI motion graphics, daily social posts, health reports from CloudWatch and production DBs.

If one function runs at agent speed and another at human speed, human speed caps everything.

What this means for you

For engineers

Your value shifts from producing code to quality of judgment. Writing code fast is worth less every month. Evaluating, critiquing, and directing AI is worth more. Can you look at a generated UI and know it's wrong before a user tells you? Can you spot the failure mode the agent missed?

For CTOs and founders

The bottom line

Harness engineering is becoming standard. Nothing in CREAO's stack is proprietary — AWS, GitHub Actions, Claude, Statsig, Linear, Sentry, Graphite. The competitive edge is the decision to redesign everything around these tools and the willingness to pay the transition cost. If you're mapping your own path, our 6-month automation-builder roadmap lays out the skills curve, and our 2-month Claude Code review covers what breaks at scale. Ready to wire up the harness? Start with the cheatsheet.

Want the full Claude Code reference? Open the cheatsheet →