Vertical AI Manifesto
The world isn’t flat, your AI shouldn't be either. A principled architecture for AI domain depth, human stewardship, and the responsibility between them
Thesis
AI investment is surging, pressure is rising, and leaders are making commitments their systems can’t support. Risk is compounding faster than governance can keep up. A major company is going to break under unstructured AI. What’s keeping it from being yours?
Most AI systems today are horizontal. They’re powerful and flexible, but not grounded. They don’t understand the domain, they don’t respect its constraints, and they don’t carry a point of view. They behave like tools without lineage, intelligence without governance.
Vertical AI is different, and not the shallow version of “vertical” that’s starting to circulate, a horizontal assistant with MCP and a memory layer stitched on. That’s a costume change, not an architecture.
Vertical AI is a discipline where an AI system operates inside a single domain, under deterministic governance, with a steward responsible for maintaining the structure the system must obey. In practice, stewardship is carried by teams, product, engineering, architects, and domain experts, who together hold the authority and responsibility for the domain.
Inside that discipline, the obligation is structural: the steward establishes the domain truth, the system must govern itself deterministically, and the domain sets the boundaries neither can violate. That’s the covenant, and there are consequences when it's broken.
If you’re reading this, you may already feel that pressure. Many are carrying it like their jobs depend on it. Leaders don’t need more promises, they need AI systems that don’t collapse under their own weight.
A system that collapses under its own weight doesn’t fail at intelligence, it fails at anatomy. The model is the heart, but the body is what keeps it upright. This is the body’s architecture:
- The Directed Acyclic Graph (DAG) is the structure.
- Research Operators act as the senses.
- The Causal Lineage stores the memory.
- The AI Airlock anchors the immune system.
- Causal Refinement Learning (CARL) provides the judgment, the frontier where this architecture learns to think.
The system is responsible for Deterministic Governance - not deterministic outputs, preserving full lineage, and enforcing its own boundaries. The human steward is responsible for maintaining the DAG as a living truth, and ensuring that new models and strategies enter through existing boundaries, evaluated, not anointed. Break the covenant on either side, and the system doesn't just degrade, it learns the wrong lessons with confidence. These are the stakes, these are the costs.
The Vertical AI Creed
Vertical AI Must:
Serve a specific domain Govern deterministically Preserve full lineage Learn from causes, not correlations Act without fear or favor Guard authority, empower imagination Honor stewardship
All Signal, No Noise
A domain is a bounded scope of authority. It has its own values, rules, workflows, safety posture, and the people who have carried its structure long before AI, the stewards. A Vertical AI system doesn’t pretend to be everything for everyone. It becomes deeply competent within a single domain, and provides a full accounting for every step of the process.
When domains interact, Vertical AI enforces cross‑domain contracts and airlocks to ensure that no domain can consume incomplete or unsafe outputs from another.
“All Signal, No Noise” means:
- no decisions without evidence
- no workflows you can’t replay
- no data you can’t trace
- no pop‑ups begging for feedback
- no hidden leaps in logic
The principles are generic, but the first domain is content, so the system learns from concrete editorial signals: edits, reader behavior, research observations, cost, safety, long‑tail performance, captured with deterministic, causal lineage.
Without Fear or Favor
I have preferences. I like certain models. I like certain strategies.
But the system doesn’t exist to validate my taste.
The system isn’t tied to a single model or a single strategy, it evaluates many, compares them, and routes to the one that performs best for the task, the tenant, and the moment. If a reasoning model outperforms a Proposer‑Critic‑Judge (PCJ) evaluation cycle, a multi‑model consensus pattern, we route to it. If a lightweight model beats a frontier model on cost‑adjusted quality, we route to it. If a tournament strategy consistently wins in counterfactual evaluations, we route to it.
Without fear or favor.
We set the values, the constraints, the coefficients, the safety rails but we don’t hand‑pick the winners. The system selects strategies and models at every step, logged, replayable, and safety-gated. Even exploration requires authorization. No black boxes. No hidden randomness. Every choice is auditable.
That’s how a system grows up.
The Spine: Steps, States, and the Directed Acyclic Graph (DAG)
Every system needs structure. Existing vertical systems have a lineage of human judgment that provides it, but that lineage needs to be made explicit, versioned, and enforceable.
For Vertical AI, that structure is the Directed Acyclic Graph (DAG).
The DAG defines which steps are legal, what context flows between them, and what the system absolutely cannot do next. This is cognition as a deterministic traversal, not a free-form agent prompt.
- Steps are the units of cognition.
- States are the validated outputs we can point to and say, “Yes, that happened.”
- And the DAG is the backbone that tells the system what can happen next and what absolutely cannot.
It’s not a workflow engine bolted onto AI. It’s the thing that keeps the whole body from flailing.
- The DAG gives the system posture.
- The Causal Graph gives it memory.
- And if we get CARL right, we give it judgment.
This is where CARL must earn its role. CARL learns over the DAG, not outside it. It updates priors for which strategies work at which steps, which models earn their cost, and how reward signals propagate across the graph. Bandits explore per step. CARL learns across steps. Both log every decision; both are replayable.
The Memory: Causal Lineage
The DAG defines what can happen. The Causal Lineage records what did.
Every step the system executes, which strategy CARL selected, what context was active, what inputs were consumed, what states were produced, is captured as an immutable event. These aren't log lines. They're structured, schema-validated records of causation: these control decisions produced these inputs, which produced these outputs.
The DAG provides the causal structure. The Causal Lineage provides the causal history. Together they make every outcome traceable, auditable, and explainable. Not through inference, but through architecture.
In practice, this is event sourcing applied to a domain-constrained topology. The hard part isn't capturing the graph, it's building the analysis layer on top: the tooling that lets CARL generate priors, compare counterfactuals, and answer "why did this work?" That layer is frontier. The graph itself is proven engineering.
Event sourcing, immutable logs, schema validation, none of this is new. The industry just forgot it applied the moment someone said 'AI.'
The Eyes: Research Operators
Most plans don’t survive first contact with reality. Most AI systems don’t even make contact.
Horizontal platforms can “browse the web,” but what they return is ephemeral: raw pages, tool outputs, and unstructured text that vanish as soon as the model moves on. Useful, but not durable.
Vertical systems need something deeper.
So ours will need eyes.
Research Operators are atomic, airlocked processes designed to turn messy reality into structured observations the system can trust. They run in isolated capsules with strict schemas, narrow tool surfaces, and single‑responsibility scopes, so every observation is validated and recorded deterministically.
They don’t just fetch information. They extract claims, evidence, sentiment, and emerging patterns, each annotated with provenance and confidence. Some frontier models can approximate this today, but the architecture makes it a first‑class responsibility rather than a prompt trick.
Perfect provenance is still an unsolved problem. No model consistently captures it without error, and we won't pretend otherwise. But the system doesn't rely on perfection. Every observation carries a provenance chain and a content fingerprint, duplicates are caught, contradictions are structurally detectable, and nothing enters the Causal Graph without a receipt. As CARL matures, it will learn which sources to trust and which to discount.
Every observation is committed into the Causal Graph as durable state, linked to the step that produced it and the decisions that consume it. That’s the difference: a Research Operator doesn’t just retrieve information; it creates lineage.
This isn’t retrieval. It’s perception with memory. It won’t just generate. It’ll investigate and remember what it learns.
The Immune System: Safety by Design
Hallucinations aren’t a defect, they’re a natural consequence of how LLMs work. Humans misremember with confidence too.
The difference is that people have judgment and governance, with guidelines and guardrails. LLMs don’t, so the system has to supply them. And if the model can’t police its own boundaries, the architecture has to. The patterns that keep AI honest aren't new. Runtime isolation, scoped capabilities, validated boundaries. A generation of platform engineers already learned this on Docker and V8 Isolates. The models are new. The discipline isn't.
The AI's mind is open, but its hands are gloved. That's the AI Airlock. Each step runs inside an Execution Capsule where the model only sees the tools that step requires, dynamically scoped when created. The model never sees the full tool surface. The Control Plane thinks. The Execution Plane acts. They never conflate. Every external signal is untrusted until it survives the Airlock, schema checks, injection detection, and human review by a domain expert. Safety is not an afterthought. It is a structural property. A stray thought can't mutate state, can't touch data, and can't impersonate truth.
Human in the Loop
Most readers already know the term, but in Vertical AI it has a specific role: the Human in the Loop is the domain expert who reviews the steps the system can’t confidently resolve on its own. They don’t govern the architecture, that’s the steward’s responsibility, but they validate the transitions that CARL or the Airlock escalate for human judgment. HITL isn’t a universal gate; it’s the system’s selective checkpoint when uncertainty, risk, or domain nuance demands a human decision.
Counterfactuals: Parallel Universes on Demand
We don't just guess the best path, we simulate the alternatives.
Through counterfactual evaluation, the system will run parallel strategies in shadow mode, testing "what if" scenarios without touching canonical state. The goal: prove that the strategy we picked wasn't just good, it was better than the paths we didn't take.
The ambition goes further: Score the Scorer. Run shadow strategies next to live strategies, record the reward model's predictions, compare the live prediction to live outcome, and use the delta to recalibrate how much the system trusts its shadow evaluations. If the reward model overvalues a signal on real data, the system corrects itself. Not just the strategy, but the judgment that selected it. That's not a small claim. It's the kind of claim you prove in public, and a future essay will do exactly that.
Organic Signals
If you need to beg for a signal, you’ve already failed.
AI evaluation is fundamentally broken. Benchmarks are static and quickly overfit. Human ratings are slow, subjective, and expensive. Reward models learn to predict raters, not real value. This pattern repeats across the industry, models score higher on tests while getting worse in practice. Vertical AI takes a different stance: the only reliable ground truth is what people actually do, not what they say or what a benchmark measures.
Vertical AI learns from organic signals — the revealed preferences embedded in real behavior. Not ratings. Not benchmarks. Not synthetic evaluations. Reality is the evaluator.
We measure:
scroll depth
dwell time
copy‑to‑clipboard
sharing and social endorsement
return visits
cohort retention
long‑tail performance
These aren’t vanity metrics. They are costly, authentic traces of human judgment. Editors provide fast, high‑resolution feedback through their revisions. Readers provide slower, deeper signals through how they move, linger, copy, share, and return.
Organic signals are self‑weighting. Quality content gets finished and shared. Safe content doesn’t get abandoned. Efficient content gets read to completion. Durable content brings people back. No single metric can dominate because all of them must remain positive simultaneously.
Causal lineage ties these signals to the exact decisions that produced them. This makes organic behavior the closest thing to causal ground truth a production system can have.
A Beginning
This architecture is built to be portable. But I'm not claiming victory before the first post ships. Content is simply where I'm starting. It's the first domain where this architecture will be tested, where the ideas in this manifesto will meet reality, learn from real signals, and show whether they actually hold up.
Some things I expect to prove:
- that a DAG as domain model guarantees deterministic governance and full lineage
- that Execution Plane isolation and AI Airlocks give Vertical AI the principled governance modern AI desperately needs
- that organic signals, captured honestly, tell us what actually worked
Some things I can't promise, but will pursue in the open:
- that CARL can learn causal structure from real editorial signals
- that scoring the scorer can earn the system trust it deserves
- that strategies can compete and improve without human hand-picking
We will be bold in experimenting. We will be transparent in outcomes. And we will be honest about our limitations.
If you follow this series of essays, you'll see the system launch and evolve in public, the parts that work, the parts that don't, and the parts that surprise us.
Your engagement, reading, lingering, sharing, or even skipping, becomes part of the feedback loop that shapes what the system becomes.
This isn’t a pitch. It’s the first step of a long walk.
The North Star
A system that amplifies human judgment rather than replacing it.
A system where every decision has a lineage, every step has a reason, and every outcome can be explained.
A system that grows through grounded signals, not noise.
Vertical AI isn’t a product category. It’s a cognitive architecture, a way of taking responsibility for intelligence. A covenant between the steward, the system, and the domain.
And this covenant belongs first to the people who have already been living it.
The ones who kept the platforms running long after the spotlight moved on. The ones who carried complexity no one else could see. The ones whose work was essential, but rarely celebrated. The ones who built with care in places where care was optional.
This manifesto is written for them, and for everyone who chooses to join them.
If these principles speak to you, stay connected.
Because this discipline grows through the people who choose to carry it:
We the builders. We the stewards. We the living.