sim-ai.org

Simulation AI Organization

VOL. 04 — NO. 01 · MARCH 2026 · EDITION 0.4.1

Foundation

sim-ai.org is a research organization studying the construction, calibration, and ethical deployment of large-scale simulation systems whose internal logic is governed by learned models. Our work occupies the seam between empirical science, computational engineering, and the humane disciplines of policy and design.

We treat a simulation as a kind of argument — a claim about the world, written in the precise grammar of computation. Like any argument, it can be honest or evasive, well-formed or sloppy, generous to its reader or hostile to scrutiny. The organization exists to articulate the standards by which such arguments may be evaluated, and to build the tools that make those standards practical.

The figure below sketches the relationship between our three founding domains — Simulation, Artificial Intelligence, and the Organizational framework that holds them in productive tension. It is not a hierarchy. It is a network of obligations.

Fig. 01.A A relational schematic of sim-ai.org's three founding domains and their secondary connectors. Edges represent obligations, not flows.

The organization was founded in 2023 by a working group of computational scientists, philosophers, and design researchers who shared a conviction: that the next generation of simulation systems would not merely model the world, but increasingly substitute for it in the deliberations of governments, firms, and individuals. Such substitution demands a discipline of its own.

We publish openly. We work in the open. We treat reproducibility not as a courtesy but as the condition of legitimacy. Every diagram in this document is constructed from primitives that any reader may inspect, modify, and rebuild.

Architecture

Our reference architecture decomposes a simulation system into four distinct stages — Ingestion, Calibration, Execution, and Adjudication — each with its own contracts, failure modes, and evaluative criteria. The decomposition is not novel; the discipline of holding to it is.

The schematic below diagrams the canonical pipeline. Boxes denote stages; arrows denote typed data flows. The annotation marks at the diagram's perimeter indicate which stage a given monitoring instrument observes.

Fig. 02.A Canonical four-stage pipeline. Adjudication observes Calibration and Execution but does not gate them in real time; its operation is asynchronous and on the public record.

Notice that Adjudication is not a gate but a witness. We have come to believe that real-time gating mechanisms — well-intentioned as they are — encourage the construction of fragile bypass logic when latency budgets are tight. A patient witness, by contrast, encourages durable hygiene.

PRINCIPLE

A simulation system that cannot be paused, inspected, and reproduced after the fact is not a research instrument. It is a rumour that happens to be implemented in software.

The execution stage is the most computationally intensive, and the most overrepresented in popular accounts of the field. Calibration is where the work actually lives. The figure below isolates the calibration substructure, showing the iterative loop between the prior distribution, the observational dataset, and the posterior under refinement.

Fig. 02.B The calibration loop. The residual node is the silent organ of self-correction; its trajectory across iterations is the most reliable diagnostic of model health.

Methodology

Our methodological commitments are few in number but binding in practice. We name them here as plainly as we can.

First: every simulation must declare its scope of validity — the conditions under which its outputs are intended to be informative. Outside that scope, the model is a piece of furniture, not a witness. We have found that the discipline of writing the scope statement before building the model is among the most clarifying intellectual exercises available to a research team.

Second: every parameter must be traceable to either an empirical measurement, a deliberate prior, or a documented assumption. The category of "what we just chose" is a category we work hard to abolish. It is, of course, never wholly abolished.

Calibration proceeds through a small number of well-understood techniques — Bayesian inference, approximate Bayesian computation, history matching for high-dimensional posteriors. We are not ideological about technique. We are ideological about transparency.

Third: every simulation we publish carries a residual report — a candid accounting of what the model gets wrong, where, and by how much. The residual report is not a confession; it is a coordinate system. Without it, claims of fit are unanchored and improvement is impossible.

These commitments are easy to list and difficult to keep. We keep them by the simple device of refusing to ship anything that violates them. The refusal is sometimes painful. It has, so far, never been regretted.

Fourth: a simulation which is consequential must be adversarially audited by parties who do not depend on its findings being correct. We maintain an internal red team and rotate external reviewers on a yearly basis. The findings of these audits are public.

Beyond these four commitments, we work in conventional ways: small teams, frequent peer review, careful version control, plain prose. The methodology is not exotic. The discipline is.