Skip to main contentSkip to footer
Beta case study

From auto-send to governed in one session

A 90-line Python script was silently sending emails to executives with no approval, no visibility, and no way to stop it. One Blueprint MCP session changed that.

Key Facts

Design documents
8 files · 183 KB, covering architecture, 52+ data models, workflows, and 30+ code examples
Senior dev work replaced
136 hours = 3.4 weeks of full-time architect work · ~$20,400 at market rate, compressed into one session
Principle compliance delta
0/10 → 10/10, all 4 Blueprint clusters covered with working, validated code
Risk reduction
70–80% of operational risks eliminated, runaway automation reduced from Critical to Low across 8 risk categories
Compliance frameworks unlocked
SOC 2, HIPAA, GDPR patterns built-in from day one, approval gates, audit trail, and state persistence included

Live validation

Two live readiness badges, before and after

Both runs are real validator outputs, not demos. The before badge links to the v1 baseline run; the after badge links to the v2 aligned run after the same-session refactor.

Blueprint Readiness Score card, after run

Validator trajectory

Two passes, every run ID public

Test #1 reached ALIGNED in two passes. Each row's run ID links to its own readiness-review page so the trajectory is independently verifiable.

PassDescriptionVerdictRun ID
v1Ungoverned baselineNEEDS_CHANGESca4b5bf6-a2c3-4223-abde-4d0ba6dc042a
v2AlignedALIGNED96a6800d-6f82-4931-9c44-5d51ec4c2217

Principle scorecard

Every flagged principle, v1 vs v2, at a glance

Four principles fired as NEEDS_CHANGES on the v1 baseline; the v2 aligned version closes all of them. The narrative below walks through each, this table is the scannable summary.

PrincipleClusterv1 baselinev2 aligned
#2, Background Work PerceptibleVisibilityNEEDS_CHANGESALIGNED
#5, Replace Magic with ModelsDelegationNEEDS_CHANGESALIGNED
#7, InspectabilityTrustNEEDS_CHANGESALIGNED
#8, Explicit Hand-offsTrustNEEDS_CHANGESALIGNED

Refactor scale

v1 ungoverned vs v2 aligned

Numbers verbatim from the in-repo BEFORE_AFTER and DOCTRINE_MAPPING source artifacts. The v2 implementation is a modular DAG with explicit approval gates and full audit trail.

Aspectv1 ungovernedv2 aligned
Lines of code~90~2,000+ (modular)
Doctrine compliance0 / 1010 / 10
User controlNonePause / resume / cancel
Approval gatesNoneBlocking, per-email
Audit trailNonePer-run, persisted
Hardcoded dataYes (recipients, prompts)No (typed inputs)
Error handlingSilent failureComprehensive, surfaced
TestabilityHard to testMocking-friendly DAG

Before / After

Before

The problem: silent, ungoverned execution

The script read documents, summarised them with an LLM, drafted follow-up emails, and sent them, automatically, without asking. No approval gate. No progress signal. No way to pause or cancel.

$ python document_processor.py
Starting document processing run...

[... 2 minutes of nothing ...]

[SENT] To: cfo@company.com | Subject: Follow-up: Q1 Budget Review
[SENT] To: csm@company.com | Subject: Follow-up: Customer Churn Analysis
[SENT] To: procurement@company.com | Subject: Follow-up: Vendor Contract Renewal

Done. Processed 3 documents and sent 3 emails.

After

After: guided, governed execution

After the MCP session, every step is visible, every email requires explicit approval, and the operator can pause, cancel, or inspect at any point.

$ python -m document_processor --request request.json

EXECUTION PLAN
══════════════════════════════════════
Documents: 3 · Stakeholders: 3
Estimated time: 2m 15s · Est. cost: $0.14

Operations:
1. Validate input (5s)
2. Summarize 3 documents (3× LLM, 25s)
3. Draft 3 emails (3× LLM, 25s)
4. Await your approval (blocking)
5. Send 3 emails (10s)

Ready to proceed? [Yes/No] yes

[14:32:04] ⏳ Summarizing 'Q1 Budget Review'...
[14:32:12] ✓ Summary: 287 tokens ($0.0009)
[14:32:13] ⏳ Drafting email to cfo@company.com...
[14:32:21] ✓ Email drafted: 156 tokens

[14:32:22] ⏸ AWAITING YOUR APPROVAL

> approve appr-001

[14:32:45] ✓ Email approved by operator
[14:32:48] ✓ Email sent (msg-20260426-001)

Validator output

What the validator found

The Blueprint MCP ran architect.validate against the original script. Four principle violations identified, each one a production risk.

#2, Background Work Perceptible

Only print() output, no persistent job status. User has no idea if work is running, stuck, or done.

#5, Replace Magic with Models

time.sleep(0.5) unexplained. Auto-send path not transparent, implicit behaviour with no system model.

#7, Inspectability

No audit trail. LLM prompts and outputs not persisted. Impossible to verify what was sent or why.

#8, Explicit Hand-offs

No approval gate. No blocking state. Emails fire to CFO, CSM, and Procurement with zero human review.

How each violation was resolved

What the validator fixed

The MCP session generated architecture documents and working code for each violation. Four principles, four targeted fixes, applied in a single session.

#2, Background Work Perceptible

Structured execution plan shown upfront with document count, time estimate, and cost. Per-operation progress logs with timestamps. Silent magic replaced with visible, predictable state.

#5, Replace Magic with Models

time.sleep(0.5) removed. Every operation declared in the execution plan before running. Auto-send replaced with an explicit --auto-approve flag; default requires human confirmation.

#7, Inspectability

Full audit trail persisted, LLM prompts, token counts, costs, message IDs, and delivery timestamps written on every run. Every decision is now traceable and replayable.

#8, Explicit Hand-offs

Blocking approval gate added before any send. Operator reviews and approves each email individually. The agent cannot proceed without explicit confirmation, no email leaves without a human decision.

Re-validation result

After the fix: architect.validate confirmed aligned

The governed implementation was re-validated in the same session. All four flagged principles resolved, status moved from Needs Changes to Aligned immediately.

Before

Needs Changes

0 / 4 principles aligned

After

Aligned

4 / 4 principles aligned

Time to fix

One session

Violations to full compliance

View the live readiness review →

Calculated ROI

Same metrics, same calculator powering every case study

Derived deterministically from this case study's profile (one session, recoverable blast radius, autonomous workflow) via /lib/case-study-roi.ts. Numbers directly comparable to the other case studies.

Senior-architect time replaced

~88 hours @ $150/hour ≈ ~$13K per agent

Production ROI per agent / year

$50K – $120K (incident prevention + audit prep + rework)

Time to identify the governance gaps

2-4 weeks of senior-architect review WITHOUT Blueprint, ~15 min / 1 validator passes WITH Blueprint

Incidents prevented (range)

3-8 per year of unintended outbound side effects (each ~4-40 hours of incident-response / rollback)

Compliance audit prep

~20-40 hours / year replaced with one audit query

Related, Pro / Teams

Run this as a Blueprint Readiness Score

The Architect Agent is the same review pattern shown in this case study, applied to your code. Call architect.validate to get a Blueprint Readiness Score (0–100, A–F) per repository, and a regression diff between runs so the next review focuses on what changed.

Sample score card

B
82/ 100

Production-ready

▲ 7

acme/customer-agent

Run your own validation

Paste your agent code or describe your workflow. The validator returns principle-by-principle findings, a readiness status, and a shareable review URL in seconds.