Three architectural papers, written in April 2026, arguing that safe autonomous AI requires three load-bearing things the current industry does not build deliberately: a structural constitution below the product, an ambient-system architecture above it, and a personal-alignment layer inside the operator running it. A fourth paper grounds the trilogy in an empirical decomposition of the codebase: where the lines of code actually live, what fraction of the tool surface shells out, how the consciousness layer breaks down by operational tier, what the audit log records. All four papers are argued from one working system — JARVIS, a local-first AI cybersecurity operations console developed by one operator on one workstation since 2026-03-16. Trilogy snapshot (Papers 1–3): 689 authored Python files, 232,458 lines of code, 27 subsystems, 81,380 hash-chained audit records as of 2026-04-18. Paper 4 uses a later snapshot (2026-04-22). A companion document carries errata, extensions, snapshot-drift across the four papers, and an "Honest Gaps" list preserved verbatim from the engineering substrate audit.
1. The Constitution Before the Product
Most AI systems are built product-first and wrapped in safety afterward. The claim of this paper is that autonomous AI systems with real capability have to be built the other way around — the safety architecture written first, as running code the product cannot bypass. The paper describes a five-layer architecture that inverts the usual arrangement (cryptographic primitives innermost, personality outermost), a serial chain of seven fail-closed gates every autonomous action must traverse, a hash-chained audit log with a documented chain break repaired and preserved, and a kill switch framed as a covenant between operator and system. One concrete case study: an April 13, 2026 trust breach where the constitution held.
→ Read paper 1 — The Constitution Before the Product
2. The Ambient Intelligence Problem
A chat assistant that forgets nothing is annoying. An ambient system that does the same is uninhabitable. The paper argues that AI systems which live alongside an operator — not inside a chat window — require memory, attention, interruption, personality, and off-mode as first-class subsystems with their own data structures, not as UX polish. Contributions include a six-layer memory hierarchy with decay constants and promotion rules, a seven-factor salience engine with an explanation surface, an interruption policy treating interruption as an operator right rather than a system default, and a personality layer realized as persistent state. Reachy is discussed as a design choice, not a requirement.
→ Read paper 2 — The Ambient Intelligence Problem
3. What Dario Didn't Say
Alignment literature treats the human in the loop as a constant. This essay argues that the human is the thing that drifts, and that personal alignment — the property of the operator-AI system considered as a whole — is a load-bearing beam of AI safety that currently nobody reinforces. It names the operator erosion dynamic (seven steps, every one lived), uses the April 13 trust breach as the paradigm case for why personal alignment matters, enumerates five things operators need (a disagreement-partner, unmediated routines, small-error practice, external accountability, willingness to turn it off), and names concrete things a lab could do. Written as an essay, not a conference paper.
→ Read paper 3 — What Dario Didn't Say
4. The 240K Decomposition
A direct empirical answer to the dismissal that local-first agent systems are "thin wrappers around CLIs." The paper audits JARVIS along four independent axes: codebase decomposition by functional bucket (the policy/safety surface is 1.1% of the tree; behavioral/personality is 6.9%; tests + GUI + scaffolding is 73.6%), tool dispatch by wrapper class (12 of 183 dispatch functions shell out, 11.2% of tool LOC; the rest is original Python), consciousness modules by operational tier (17 of 35 reach the operator on a live turn; 11 accumulate ambient state; 7 are inert or conditionally gated), and the 87,315-row audit log by event type (78.7% concentrated in three event types; autonomy decisions sum to 5.5%). Across all four axes the scaffolding-to-decision ratio sits in the same order of magnitude (~9:1 to ~14:1). The paper's claim is narrow: agent systems at scale decompose into scaffolding surrounded by small moments of decision, not model calls surrounded by tools — and this contradicts the wrapper framing at every axis.
→ Read paper 4 — The 240K Decomposition
How to read them
The papers are independent. Each argues its own claim and can be read alone. If you want them in order: Paper 1 is the substrate, Paper 2 is what gets built on top of it, Paper 3 is about the human running the result, Paper 4 is the empirical floor under all three. If you are arriving from an AI safety or alignment background, Paper 3 is probably the one that opens the conversation most directly. If you are arriving from a systems or security background, Paper 1 is. If you are arriving from HCI or ambient computing, Paper 2 is. If you want the empirical answer to "is this real engineering or a thin wrapper?", Paper 4 is. The companion document is for readers who want errata, antecedent literature pointers (Bainbridge, Hollnagel), and the "Honest Gaps" list — everywhere earlier claims about the system did not match code reality.
The author is a twenty-two-year-old self-taught builder in San Diego. The papers are offered in that spirit — one operator, one machine, working out loud. Corrections welcome.
System facts — see SYSTEM_STATE.md for current counts
Numbers below are a 2026-04-18 snapshot preserved as a dated artifact. For current counts, see the live system state.
The figures below are the v1.0 canonical numbers for Papers 1, 2, and 3, verified by static inspection against the live JARVIS source tree and databases on 2026-04-18. Every number quoted in those three papers matches this block.
| Fact | Value |
|---|---|
| Authored Python files | 689 |
| Lines of authored code | 232,458 |
| Subsystems | 27 |
| Registered tools | 260 |
| Hash-chained audit records | 81,380 |
| Brain-graph nodes / edges | 4,681 / 4,337 |
| Days of continuous development | 33 (since 2026-03-16) |
Total repo (including vendored cb-mpc and tests): 731 files, 241,487 lines. The figures above are the authored core — JARVIS is 689 Python files Ali wrote by hand.
Paper 4 uses a later snapshot (2026-04-22: 766 files, 248,809 lines, 87,315 audit rows). The companion document carries the full snapshot-drift table across all three dated audits.