Living Models - Table of Contents
Causal Intelligence for the Decisions That Actually Matter
The Living Models framework didn’t start as a book proposal or a research agenda. It started with an email.
Someone reached out with a concrete problem: how do you build intelligence that actually keeps pace with decisions that have to be made in fast-moving environments? The existing options — legacy analyst reports, dashboards, predictive models — all share the same fundamental flaw. They’re built to describe the past or extrapolate from it. They have no mechanism for reasoning about what happens when you deliberately change something.
That email prompted a first-principles question: what would it actually mean to build something genuinely different? Not a faster dashboard. Not a more frequently updated report. Something categorically different in how it reasons.
The answer that emerged is what I’ve been calling a Living Model — causal, counterfactual, continually updated, and organized around actionable interventions rather than descriptions. The core idea is that the gap between what analytics systems are built to do and what strategic intelligence actually requires is large, consequential, and now closeable. Closing it requires moving up Pearl’s ladder from association to intervention — and that’s not an incremental improvement. It’s a different kind of reasoning entirely.
I write about nascent ideas in public because that’s how I make them better. Private thinking produces private blind spots. Writing forces precision. Readers push back. The argument gets stronger or it gets abandoned. The hypothetical.ai Substack exists for exactly that reason — to work the math and the architecture out loud before anything gets built.
The data was never the problem. It was always the question.
Living Models
Causal Intelligence for the Decisions That Actually Matter
Nik Bear Brown
The data was never the problem. It was always the question.
Preface
The Monday Morning Meeting Why dashboards are reliable for understanding the past — what happened, what was spent, what was sold — but unreliable for understanding what is going to happen, particularly as lifecycles rapidly compress. Using dashboards to navigate the road ahead is driving via the rear-view mirror. What this book is, who it is for, and what it asks of the reader. A note on the relationship between theory and practice — and why this book refuses to separate them.
Part One: The Problem
Why three decades of analytics have produced sophisticated hindsight and very little foresight
Chapter 1: The Dashboard That Lied The anatomy of a descriptive analytics failure. What correlation can tell you, what it cannot, and the organizational cost of not knowing the difference. The four rungs of analytics maturity — and why most enterprises are stuck on rung one.
Chapter 2: The Map That Doesn’t Move Why predictive models fail under intervention. The observational distribution versus the interventional distribution. What happens when strategy — the deliberate act of making the future different from the past — meets a model trained on the past. The Innovator’s Dilemma as a failure of organizational incentives and time horizons, not merely a strategic one: executives often see the disruptive threat but are incentivized to downplay it, protecting the status quo while the long-term horizon — itself compressing — goes unmodeled.
Chapter 3: What We Mean When We Say “Real-Time” The abuse of “real-time” in enterprise software marketing. The difference between data latency and model latency. Why continually updated is the honest term. What it actually requires to keep a model current — technically, organizationally, and epistemically.
Chapter 4: Risk Is Two Numbers, Not One The collapse of probability and impact into a single risk score — and what it costs. Expected Value of Intervention as the foundational metric. Why a five percent chance of catastrophe and a fifty percent chance of inconvenience require different decisions. How almost every risk framework in commercial use gets this wrong.
Part Two: The Theory
The mathematical foundations of causal intelligence — made readable
Chapter 5: Pearl’s Ladder Judea Pearl and the three levels of causal reasoning: association, intervention, counterfactual. The do-operator and what it represents. Why the move from observational to interventional reasoning is not an incremental improvement but a categorical one. A working guide to the Ladder for non-mathematicians.
Chapter 6: Graphs That Think Directed Acyclic Graphs as maps of mechanism. Structural Causal Models and what they encode. The difference between a regression equation and a causal graph. How the same data can be consistent with multiple causal structures — and why this is a mathematical result, not an algorithmic limitation.
Chapter 7: The Equivalence Problem Markov equivalence classes and why data alone cannot orient every edge in a causal graph. The Completed Partially Directed Acyclic Graph. What this means for automated causal discovery — and why human domain knowledge is not a convenience but a mathematical necessity. How to resolve equivalence through interventional reasoning rather than more data.
Chapter 8: Estimating Effects From graph structure to quantified causal effects. The backdoor criterion and why standard regression estimates are frequently biased. Double Machine Learning and what it corrects for. Susan Athey’s causal forests and the estimation of heterogeneous treatment effects. Why the average effect is rarely the decision-relevant fact.
Chapter 9: The Counterfactual Pearl’s abduction-action-prediction procedure. Pre-factual simulation versus retrospective counterfactual analysis. How to reason about a world that never happened. The individual-level counterfactual and why it is the hardest and most valuable form of causal reasoning. Clinical precedents and their organizational analogs.
Chapter 10: Confounders, Colliders, and the Limits of Observational Data The unconfoundedness assumption and when it fails. Latent confounders — the competitor’s internal meeting, the macro-sentiment shift, the organizational change that preceded the attrition spike. Collider bias and Berkson’s paradox. What sensitivity analysis can tell you about how wrong your model might be. The honest account of what causal inference cannot do.
Chapter 11: Treatments Randomized controlled trials and why they remain the gold standard. Esther Duflo’s experimental design at scale. The translation of clinical “treatment” into organizational “intervention.” Stable Unit Treatment Value Assumption and when network interference breaks it. Spillover effects, herd immunity, and the social systems that violate SUTVA by design.
Chapter 12: The Plumber’s Objection Duflo’s “Economist as Plumber” and what it means for causal AI. The distance between a correct causal estimate and an effective organizational change — and what fills it. Why models provide very little guidance on which implementation details will matter. The Ne-FMS case and what fixing the plumbing actually looks like.
Part Three: The Architecture
How to build systems that actually use this theory
Chapter 13: The Living Model Defined The four properties that define a Living Model: causal, counterfactual, continually updated, treatment-oriented. How these properties distinguish a Living Model from a dashboard, a predictive model, a digital twin, and an ontological system. The analytics maturity table revisited. What “orchestrated outcomes” actually means.
Chapter 14: The Expert in the Room Why causal graphs cannot be built from data alone — and why this is a mathematical result, not a limitation of current tools. The knowledge bottleneck: the gap between what a domain expert knows implicitly and what a causal model requires explicitly. The specific discipline that has been working on this problem for two decades is Knowledge Engineering with Bayesian Networks — a field with rigorous methods for structured elicitation, conditional probability assessment, and graph construction from expert judgment. It has produced validated protocols used in medical diagnosis, military intelligence, and environmental risk assessment. It has not reached the corporate boardroom. This chapter explains what the field knows — variable identification, edge elicitation, consistency checking, confidence calibration — and why the gap between research and practice persists: the methods are slow, require trained facilitators, and produce outputs that don’t fit neatly into existing strategy or analytics workflows. The Living Model architecture in Part Three is, in part, an answer to that gap.
Chapter 15: How Experts Get Causation Wrong The systematic cognitive biases in expert causal reasoning: collider blindness, feedback loop simplification, domain-matching heuristics. Berkson’s bias as the canonical illustration. Why “more covariates” is not always better. What the research says about when to trust expert causal judgment and when to interrogate it.
Chapter 16: The Machine That Interviews the Expert The model for this chapter already exists — not in causal AI research, but in brand strategy. LLM-guided interview systems like the Nina framework demonstrate that a well-designed system prompt can do what a skilled human interviewer does: ask one question at a time, refuse to proceed until the answer is sufficient, hold prior answers in context, surface contradictions, and progressively build a structured output from unstructured expert knowledge. Nina does this for brand identity — moving from intake through archetype to creative brief through a disciplined sequence that cannot be shortcut. The causal elicitation system described in this chapter applies the same architecture to a harder problem: building a first-pass causal graph from a domain expert who has never heard of a DAG.
The chapter covers the full architecture: variable confirmation (what are the things that matter in this system, and are they measurable?), edge elicitation (does X cause Y, or does Y cause X, or does something else cause both?), interventional disambiguation (if we changed X deliberately, what would happen to Y — and is that different from what happens when X changes on its own?), and confidence calibration (how certain are you, and what would change your mind?). Each stage maps directly to a phase in the Nina intake sequence — the chapter makes this analogy explicit, using it to show executives a system they can already picture before introducing the causal machinery underneath.
The minimum viable interview: forty-five minutes to a first-pass causal graph, suitable for handoff to automated discovery algorithms. Multi-agent design and how different reasoning modes — one agent eliciting, one checking consistency, one flagging equivalence ambiguities — divide the work that a single interviewer cannot reliably do alone. What CausalChat-class implementations have demonstrated in practice, where they stop short, and what the Nina parallel reveals about the design principles that make the difference between an interview that extracts knowledge and one that merely confirms what the expert already planned to say.
Chapter 17: Resolving the Graph From expert-provided skeleton to fully oriented DAG. How automated discovery algorithms — PC, GES, NOTEARS, FCI — refine expert-provided structure. The CPDAG handoff and what it requires from the expert interview output. When to run more data collection and when to run more expert sessions. The validated graph as a living artifact, not a finished product.
Chapter 18: From Graph to Decision Parameterizing the graph — estimating conditional distributions from data once structure is fixed. Running the counterfactual: abduction, action, prediction in a business context. Ranking interventions by Expected Value. The constrained knapsack — translating ranked interventions into portfolio decisions under resource constraints. What the output looks like to a strategy executive.
Chapter 19: The Causal Brain Executive Report A Living Model produces one output that matters to an executive: a ranked recommendation with the evidence that supports it, the assumptions that could break it, and the counterfactual that justifies acting now rather than waiting. This chapter defines what that report contains and why each element is there — not to explain the model, but to make the recommendation auditable by someone who will never see the model at all.
The report structure follows a deliberate logic. The recommendation comes first — specific, ranked, owned. Not “consider these options” but “do this, because the model estimates this intervention produces the highest Expected Value under current conditions.” The evidence section follows: which causal variables drove the recommendation, how confident the model is in each edge, and where the graph is thin — the nodes where expert elicitation was the only data source and observational data has not yet confirmed the structure. The assumptions section names what would have to be true for the recommendation to be wrong — not as a hedge, but as an audit trail. And the counterfactual closes it: what the trajectory looks like if the recommendation is not taken, and at what point the next-best intervention becomes more valuable than the recommended one.
LLM narration is the mechanism that produces this report from model outputs a board cannot read directly. The chapter covers what that narration must do — translate intervention rankings into plain-language recommendations, surface confidence levels without false precision, flag structural uncertainty without undermining the recommendation — and what it must never do: explain the model, show the graph, or present options as equally valid when the model has ranked them.
Visualization serves the evidence layer, not the recommendation. The chapter covers the specific tradeoff every causal visualization faces: a full causal graph is structurally honest but illegible to most executives, and most simplifications that make it legible destroy the structural honesty. The resolution is not a better visualization of the graph — it is a visualization of the decision, with the graph available as an appendix for those who want to interrogate it.
The chapter closes with the accountability question this project raises directly: if the model recommends and the executive decides, where does responsibility sit when the decision is wrong? The answer is not in the model. It is in the report — the auditable record of what the model said, what evidence supported it, and what the decision-maker chose to do with it.
Chapter 20: Keeping the Model Alive Bayesian updating of edge parameters as new data arrives. Structural change detection — when a shift in the data requires revisiting the graph, not just the parameters. Model drift in causal systems and how to detect it. DecisionOps versus MLOps — tracking decision ROI, not model accuracy. The minimum viable feedback loop for an organization without a data science team.
Part Four: The Frameworks
Christensen, Damodaran, and the theory-guided Living Model
[CS: These are two of the better-known business academics with established frameworks. Christensen is widely recognized; the Christensen Institute is a potential collaborator. Damodaran is a finance professor (not strategy) known personally from NYU — generous with his IP, potentially open to working with us. These two are a starting point, not the limit. Other academic frameworks should be incorporated, including from NEU faculty. To be discussed.]
Chapter 21: Frameworks Are Not Models Why Christensen and Damodaran are inputs to causal models, not causal models themselves. The correct role of theoretical frameworks: feature engineering, Bayesian priors, anomaly flags, DAG scaffolding. What “theory-guided AI” means in practice — and why it is categorically different from either pure data-driven discovery or pure framework application. A preview of the structure that follows: for each framework, the book moves in three steps — what the framework actually argues, where its causal structure is hidden or incomplete, and what a Living Model built on that foundation can do that the framework alone cannot.
Section A: Christensen
Chapter 22: What Christensen Actually Argued The precise claim of disruption theory — not the popular misreading of it. Low-end and non-consumption entry. Why the entrant’s inferiority is a structural feature, not a weakness. The performance trajectory dynamic and why it is nonlinear. The rational trap: why the incumbent’s best response accelerates its own displacement. What the framework explains well, what it explains poorly, and the three things it cannot tell you at all — timing, mechanism, and counterfactual. Why most companies that have “read Christensen” still get disrupted: the gap between pattern recognition and causal understanding.
Chapter 23: The Disruptive Innovation DAG Building a causal graph from Christensen’s disruption theory. The structural equations behind low-end entry, performance trajectory, and mainstream displacement. Upstream drivers, midstream mediators, downstream outcomes — and the feedback loop that makes disruption self-reinforcing once started. What real-time signals — competitor pricing, job postings, user reviews, patent filings — look like as inputs to a disruption-theoretic causal model. The signal integration problem: why each signal is ambiguous alone and how the DAG resolves the ambiguity. Where Christensen’s framework breaks — the incumbent response problem, the platform shift confounder, the hindsight bias embedded in the theory’s most famous cases.
Chapter 24: The Disruption Audit — A Case Study A single incumbent/entrant pair carried from first signal to displacement confirmation. What the leading indicators showed, when they showed it, and what the incumbent said at the same moment. The DAG populated with real data. Intervention ranking at three points in the timeline: when the model would have recommended a separate unit, when it would have recommended acquisition, when it would have recommended managed retreat. The counterfactual: what would the trajectory have looked like under the highest-ranked intervention at the earliest detection point. What this case demonstrates that Christensen’s framework, applied conventionally, cannot.
Section B: Damodaran
Chapter 25: What Damodaran Actually Argued Damodaran is a finance professor, not a strategy theorist — and that distinction matters for how his frameworks should be used. His corporate life cycle framework: the five stages, the financial signatures of each, the logic of value creation when returns exceed cost of capital and destruction when they don’t. His equity risk premium work: why historical premiums are backward-looking and biased, what the implied ERP measures instead, and what it still gets wrong — risk as a single number, the efficiency assumption doing load-bearing work in the background, the collapse of probability and impact that Chapter 4 identified as the foundational error. What Damodaran gives the practitioner that almost no one else does: rigorous, freely available, annually updated tools. What those tools cannot do: model the mechanism before the market sees it.
Chapter 26: The Life Cycle as Causal Structure Damodaran’s corporate life cycle encoded as a DAG. The financial signatures of each stage — reinvestment rate, margin trajectory, capital allocation — as candidate causal variables. The transition problem: why late Mature Growth and early Mature Stable are nearly identical in levels but structurally different in second derivatives. The slow edges hardest to detect: organizational capability atrophy, the narrowing ROIC/WACC spread, the reinvestment rate inflection. How a Living Model monitors the transition from Mature Growth to Mature Stable. Predictive intervention before the decline phase begins. The ERP connection: why the implied equity risk premium will not move until the market sees what the causal model already shows.
Chapter 27: The Decline Inflection — A Case Study A single public company carried through the Mature Growth to Decline transition. What the dashboard showed at the transition point — and why a reasonable analyst would have seen nothing alarming. What the causal signals showed: the second derivatives, the organizational atrophy indicators, the management narrative gap between stated strategy and actual capital allocation. The DAG populated with eight quarters of financial trajectory data. The counterfactual: what the model would have recommended at the highest-leverage decision point, and what the trajectory would have looked like under that intervention. The lag between earliest causal signal and market price response — and what that lag means for the implied ERP as a forward-looking tool.
Section C: The Collision
Chapter 28: The Collision Model The SaaS Margin Collision as a worked example of what happens when Christensen’s disruption dynamic and Damodaran’s life cycle transition operate simultaneously. Compute cost, labor elasticity, and pricing architecture as a three-variable causal system. Non-linear ripple effects versus additive forecasting. The Life Cycle Compression Index as a Living Model output. What the model tells a CEO that a static forecast cannot — and what it tells a board that neither Christensen nor Damodaran, applied separately, would have surfaced.
Part Five: The Cases
Living Models applied to real strategic decisions
Chapter 29: The Pricing Reset A seat-based SaaS company facing the agent disruption scenario. The causal model of pricing architecture, customer retention, and competitive entry. Intervention ranking under a resource constraint. The counterfactual: what would revenue look like if the pricing model had shifted two years earlier?
Chapter 30: The Supply Chain That Broke A manufacturing firm, a tariff shock, and a causal model built on publicly available data. SUTVA violations in supplier networks. How the Living Model would have ranked contingency interventions before the disruption. What actually happened and what the counterfactual suggests.
Part Six: The Frontier
What Living Models cannot yet do — and what comes next
Chapter 31: The Latent Confounder Problem The hardest unsolved problem in applied causal inference. What sensitivity analysis can and cannot tell you. Methods for partial identification under unmeasured confounding. The honest account of where Living Models fail and why that failure is informative rather than disqualifying.
Chapter 32: Networks and Interference SUTVA violations in social and market systems. Bipartite experiments, network unconfoundedness, and spillover effect estimation. What Living Models look like in markets where every unit’s outcome depends on every other unit’s treatment. The herd immunity analogy and its organizational equivalents.
Chapter 33: The Agentic Living Model From decision support to autonomous decision execution. The architecture of systems that not only recommend interventions but implement them. The governance problem — what “Urban Reasonableness” means for autonomous causal agents. The EU AI Act and what regulatory compliance requires from Living Model architecture.
Chapter 34: Causal Digital Twins The difference between Palantir’s ontological modeling and genuine causal simulation. What a Causal Digital Twin actually is — SCMs with real-time sensor fusion, automated discovery, and counterfactual generation at scale. Where this technology stands, what it requires, and what it will make possible when it arrives.
Appendices
Appendix A: A Glossary of Causal Terms for Strategy Executives DAG, SCM, do-operator, CPDAG, CATE, ATE, backdoor criterion, collider, confounder, Markov equivalence, counterfactual, pre-factual, interventional distribution — defined in plain language with organizational examples.
Appendix B: Pearl’s Do-Calculus — The Three Rules The mathematical foundation, made readable. What each rule permits, what each rule requires, and a worked example in a business context.
Appendix C: The Minimum Viable Interview Protocol The forty-five-minute structured elicitation session — question by question. What each question is designed to extract. How to handle an expert who collapses levels. The output format for handoff to automated discovery.
Appendix D: Software and Tools Current state of the causal AI ecosystem: causaLens/decisionOS, DoWhy, EconML, CausalML, NOTEARS, CausalNex, Bayesia, Netica. What each does well, what each requires, where each stops short. A practical guide for the organization starting to build.
Appendix E: The Living Model Reading List Pearl’s The Book of Why and Causality. Athey and Wager on causal forests. Duflo’s Poor Economics and the Economist as Plumber lecture. Christensen’s The Innovator’s Dilemma. Damodaran’s The Corporate Life Cycle. The academic papers behind the commercial platforms. Annotated for the reader who wants to go deeper.
hypothetical.ai — causal intelligence for executives who actually have to make decisions

