Skip to main content

Salesforce

The Architect’s Survival Guide to AI-Augmented Delivery

What changed at TDX 2026, and what architects need to do about it.

9 min read
By Anna Bromley
Architecting AI session slide at TDX 2026: AI suggests, architects decide.

The first time you watch AI rewrite a chunk of your codebase, you feel a small flicker of panic. You asked it to refactor one function. It quietly touched seven. You click through the diff and the floor moves slightly under your feet. You either accept the change because the build still passes, or you start scrolling line by line trying to work out what just happened to the thing you wrote last week.

If you’ve vibe coded for any length of time, you know the feeling.

At TDX 2026 the Architecting AI session put a number on it. One AI cycle, the speakers said, took a codebase from 8,900 lines down to 4,000. Build still ran. Tests still passed. Nothing screamed. The AI’s summary said the change was small. The code said otherwise.

That’s the gap this article is about. Not whether AI is useful in delivery; it obviously is. The question is whether the architect’s role has caught up with what the tooling now does without asking. Below: what changed at TDX, the new disciplines worth adopting on Monday, and the parts of the SDLC where you cannot delegate, no matter how good the model gets.

The job changed. Most architects haven’t.

Two years ago, an architect’s value was visible in what they shipped. Diagrams, schemas, code reviews, the integration patterns that survived audit. The model worked because the architect was the one closest to the build.

The Architecting AI session at TDX put it differently. You’re no longer just a builder. You’re the one guiding the builder. The role inverts. The architect’s value is in what they prevent the AI from getting wrong, not in the volume of what they personally produce.

AI suggests. Architects decide.

This is uncomfortable for two reasons. First, it’s harder to measure. A well-judged “no” doesn’t show up in a sprint report. Second, the AI is fast and convincing. As one of the speakers put it: “It just loves to stroke you. ‘You have a beautiful, complete solution.’ And we’re like... calm down, there’s more to go.”

Architects who stay measured by velocity will lose to AI on velocity, every time. Architects who reposition around the parts of the SDLC the AI cannot judge become more valuable. Those parts: risk, intent, integrity, regulatory posture. Your job now is to stay in command, and to make the case for staying in command.

Two disciplines, two risk profiles

Most teams treat AI as one thing. It isn’t. There are two distinct disciplines, and conflating them is what gets architectures into trouble.

Using AI means consuming AI tools to produce code, designs and tests faster. The risk profile is the architect’s: drift, silent regression, designs that look complete but aren’t.

Building AI means architecting a system that itself reasons, acts, and remembers. The risk profile is the customer’s: unpredictable behaviour, runaway tool calls, audit failure when something goes wrong.

They share a model and not much else. The governance, the controls, the audit trail: different for each.

Two disciplines, two risk profiles, two audit trails.
Using AIBuilding AI
GoalProducing artefacts faster: code, designs, testsProducing a system that reasons, acts, and remembers
Primary riskDrift between intent and codeDrift between intent and behaviour at run time
Where humans stay in commandDesign checkpoints, diff review, TRD lockTool boundaries, kill switch, observability, blast radius
Audit trail you needCommit-per-cycle, TRD-versioned changesDecision logs, agent traces, tool invocation history

The AI-augmented SDLC: where the checkpoints live

Walk it phase by phase. At every boundary, the architect either holds a checkpoint or loses one.

Architecting AI session slide: Plan phase, architect involvement upfront.

Plan phase: architects belong upfront, not at sign-off.

Plan

Architects belong upfront, not at sign-off. A plan with no architectural input is a plan that becomes expensive to fix at design phase. Lock the scope, the constraints and the non-functionals before any AI cycle starts.

Requirements

AI inflates. One sentence from a stakeholder (“I need a system to track contractor pay”) becomes two pages of requirements, one hundred pages of use cases, and ten thousand lines of code. That’s the balloon problem. The checkpoint is the Technical Requirements Document (TRD): locked and versioned, anchoring every cycle that follows.

Design

Fast output, weak architecture is the failure mode. AI produces designs that look finished. The diagrams are there, the sequence flows, the schemas. The cross-cutting concerns are not. Identity, observability, blast radius, multi-tenancy, regulatory posture. None of those appear unless you prompt for them. This is where the architect’s special sauce lives.

Architecting AI session slide: Design phase risk, fast output but weak architecture.
Architecting AI session slide: Design phase special sauce, the architect's irreducible value.

Build

Vibe development is fine when it’s bounded. The discipline is commit-per-AI-cycle: one feature, one cycle, one commit, diff-reviewed before merge. (More on this below.)

Architecting AI session slide: Build phase, vibe development bounded by architectural discipline.

Vibe development is the start, not the strategy.

Test

AI-generated tests prove behaviour, not intent. A passing suite is not a guarantee that the system does the right thing. Humans validate intent. The architect’s job is to define what “right” means before the build, not after the bug.

Deploy

Borrow Google’s playbook from the Agentforce Journey session: shadow first, then supervised, then automated, only at 80–90% accuracy thresholds. Treat rollout as part of the architecture, not the project plan.

A design can look complete and still be architecturally incomplete

This was the most-quoted sentence of the Architecting AI session, and it’s the one to put on the wall.

AI is excellent at producing artefacts that look finished. The diagram has all the boxes. The schema has all the fields. The sequence flow has all the arrows. What it misses is the design behind the design. Where does identity flow? What’s the blast radius if this service fails? Where does observability live? Who owns the data model when two products diverge?

Two parts of an AI-augmented design stay human-only.

Do not trust AI on the data model.

It’s a hard line. Schemas drive everything downstream: refactor cost, regulatory exposure, the integrity of every report a CXO ever sees. AI will produce a schema. Don’t ship it without an architect having designed it.

The second part is integration boundaries. The Agentic Architecture session at TDX framed enterprise architecture as urban planning. Without it, you don’t get a city; you get a shanty town. AI will build neighbourhoods enthusiastically. Architects zone the city. Decide where APIs go, where MCP fits, when A2A is appropriate, when bridges make sense. AI does not have the context to answer those questions, won’t, and the cost of getting them wrong is structural.

Agentic Architecture session slide at TDX 2026 framing enterprise architecture as urban planning.

Enterprise architecture is urban planning. Without it, you get slums.

The architect’s real product is the design behind the design.

Commit-per-AI-cycle: the discipline that saves your codebase

Back to the 8,900-line story. The reason it didn’t scream is that there was no commit boundary between the version the team trusted and the version the AI rewrote. The AI’s summary was reassuring; the diff would have been alarming. Nobody opened the diff.

Commit-per-AI-cycle is the discipline that fixes this. Four rules:

  1. One feature, one cycle, one commit. If a cycle changes more than one thing, split it.
  2. Diff-review every cycle. Read the diff, not the AI’s summary of the diff.
  3. Lock the TRD before each cycle. Reference it explicitly in the prompt. Without it, instructions are slippery; they drift across cycles and you can’t tell what the model is optimising for.
  4. Roll back without ceremony. A cycle is cheap. Rolling back isn’t failure. It’s how the workflow works.

One Architecting AI quote captures the upside: “I could erase all my code, say ‘build it according to the TRD,’ and it would almost appear correct.” That’s the test. If your TRD is good enough that the codebase could be regenerated from it, the architecture is in shape. If it isn’t, the architecture is hidden in the heads of whoever last committed.

The architect’s survival checklist

Print this. Tape it to your monitor. The disciplines below survived contact with the worst rooms at TDX.

Eight disciplines for AI-augmented delivery, and what each one protects.
DisciplineWhat it looks likeWhat you protect
Lock the TRD before generationTRD signed off and version-controlled before any AI cycle startsDrift between intent and code
Own the data modelSchema designed by humans, reviewed against business capabilitiesRefactor cost and regulatory exposure
Commit-per-cycleOne scoped change per AI cycle, diff-reviewed before mergeSilent regression
Architect at design checkpointCross-cutting concerns reviewed and signed off before buildArchitectural incompleteness
Verify intent, not just behaviourTests prove the system does the right thing, beyond simply runningConfidently wrong outputs
Bound the agent's blast radiusTool list scoped to intent, with a kill switch and a budget capRunaway cost and audit failure
Observability before deployLogs, traces and agent decisions persisted from day oneDebuggability and trust
Phase the rolloutShadow, then supervised, then automated, with explicit accuracy thresholdsProduction safety

What this means for your CTO, CIO or COO

If you’re an architect, you can stop here. The next two paragraphs are the version to forward up the chain.

The shift is this. AI is now a delivery accelerator, but only if the architectural disciplines around it are funded and respected. Cut those, and the productivity gain disappears into rework, audit risk, and the kind of silent regression that doesn’t show up until a regulator or a customer finds it for you. Your architects are not the bottleneck. They are the audit trail.

The control to ask for: every AI cycle reviewed, every architectural decision logged, every rollout phased. The clarity to ask for: a TRD per stream that’s good enough to regenerate the system from. The confidence to ask for: a delivery you’d put in front of a regulator on Monday.

Want a calm review of your AI-augmented delivery?

If your delivery is moving faster than your architecture can defend, that’s the gap we close. Start with a Delivery Control Snapshot: a one-page risk view of your AI-augmented programme that your sponsor can act on this week.

Get in touch

Further reading

Source material and authoritative references for the topics in this article.

About the author

Anna Bromley - Director, Agile Delivery

Anna Bromley

Director, Agile Delivery

Connect with Anna

You may also like…

Make transformation deliver

Get in touch