We Pointed It at Ourselves

We built 1.1 million lines of production code with AI assistance. Health checks said HEALTHY. Tests passed. Then we built a behavioral verification system — turned it on our own codebase — and found what every team using AI coding agents should be looking for.

What We Believed Before We Checked

We had 1.1 million lines of production code. All built with AI assistance across multiple providers.

We had 18,000+ test functions. A 98/100 self-audit score. 57,000+ evolved patterns in our learning system. Health endpoints returning HEALTHY. Dashboards green across the board.

By every conventional measure, the system was working. Tests passed. Code compiled. Services responded. We had more coverage metrics than most teams dream of.

Then we asked a different question: does data actually flow through the expected paths? Not 'does the code exist?' Not 'does the endpoint respond?' But: when real data enters Pipeline A, does it come out the other end with a real result?

The answer changed everything we thought we knew about AI-assisted development.

Proven Across Two Complete Builds

The Discovery

When we ran behavioral verification against our own codebase, the findings were devastating. What tests and health checks called 'working' was, in many cases, structurally connected but behaviorally dead.

  • 3 Completely silent data pipelines
  • 99 Calibration results — all hardcoded defaults
  • 1,218 Evolution cycles with zero diversity
  • 82 Runtime violations invisible to testing

The Fix

With behavioral verification identifying the actual problems, the fixes were targeted and fast. No more guessing. No more whack-a-mole debugging. Every issue had a clear root cause and a verifiable fix.

  • 112 Issues fixed in one sprint
  • 100% Flow verification after remediation
  • 4 Failure types classified and catalogued
  • 0 Silent wiring remaining after verification

What Behavioral Verification Revealed

Silent Pipelines

  • Dead data pipelines discovered 3
  • Time pipelines were silent before detection Months
  • Health check status during silence HEALTHY
  • Tests passing during silence All

Behavioral Violations

  • Runtime violations caught 82
  • Calibration results that were hardcoded defaults 99
  • Evolution cycles with zero diversity 1,218
  • Violations visible to conventional testing 0

Remediation Results

  • Issues fixed in one sprint 112
  • Declared paths verified end-to-end 100%
  • Failure types classified 4
  • Recurring issues after behavioral gates 0

System Scale

  • Total lines of production code 1.1M+
  • Evolved patterns in learning system 57,000+
  • Test functions 18,000+
  • Self-audit score 98/100

8 patents protecting the methodology

PAT-001
Constitutional Capture

Infrastructure-level recording that can't be bypassed

PAT-002
Self-Calibrating Confidence

Pre-execution risk scoring that improves with outcomes

PAT-003
Cross-Model Transfer

Accumulated intelligence that survives provider switches

PAT-004
Immutable Audit Infrastructure

Event-sourced trails for regulatory compliance

PAT-005
Pattern Evolution

Genetic evolution of operational patterns

PAT-006
Self-Verifying Audit

System that validates its own integrity

PAT-007
Multi-Model Consensus

Decision-making across providers with evolving weights

PAT-008
External Safety Coordination

Provider-independent oversight architecture

Blind evolution produced production-quality patterns

The genetic algorithm's domain-specific transformation operators were robust enough to surface viable patterns through structural validity alone — without outcome feedback.

When the outcome feedback loop was connected, improvement potential is estimated at 2–5x.

Architecture so robust it produced results even with an incomplete feedback loop.

33 years of compounding expertise

1993 First learning systems
1996 x.hlp Technologies founded
2004 x.hlp acquired (now inside SAP)
2005–2019 Energy sector — grid analytics
2020 CleanAim founded
2024–2025 1.1M lines built, Silent Wiring discovered
2025 8 patents filed
2026 Silent Wiring book published, EU AI Act enforcement

Explore the Evidence

If a team that built the verification system still found 3 silent pipelines in their own code — what does that tell us about the state of AI-generated code everywhere?

Join the Waitlist