Tiers of Abstraction — Zero-Tier, Tier 1, Tier 2
TL;DREvery principle in this scaffold assumes something about the substrate it applies on. The cleanest honesty-check is the paper-and-pencil test: would this still be true if you replaced all technology with notebooks and filing cabinets? If yes, it’s zero-tier. If it depends on digital storage, it’s tier 1. If it depends on LLMs or agentic architecture, it’s tier 2.
The claim
Section titled “The claim”Principles come in scopes of applicability. A claim like “attention is finite” holds for any cognizer — human, AI, or a committee — regardless of whether they use paper, filesystems, or neural interfaces. A claim like “files should have canonical paths” assumes you have files. A claim like “skills load into agent context” assumes LLM architecture.
These are different scopes, and treating them as equally solid is a category error. Our 01-kernel/principles/ folder currently mixes all three scopes without flagging which is which. This page names the distinction.
The three tiers
Section titled “The three tiers”Each tier builds on the previous, narrowing scope from universal cognition to LLM-specific tooling:
flowchart TD A["Tier 0 — Any Cognizer<br/>Paper + Pencil<br/><br/>Attention finite<br/>Capture → Work → Output<br/>Temperature gradient<br/>Convention as compressed decision"]
B["Tier 1 — Digital Technology<br/>Filesystems + Networks<br/><br/>Single canonical addressability<br/>Progressive disclosure<br/>+ all of Tier 0"]
C["Tier 2 — Agentic / LLM-Specific<br/>Context windows<br/>Agent architectures<br/><br/>Skills vs Agents<br/>Context engineering<br/>+ all above"]
A --> B B --> C
style A fill:#c8e6c9 style B fill:#bbdefb style C fill:#ffe0b2The cards below link into each tier’s detail section:
Zero-tier — Knowledge work, any substrate
Section titled “Zero-tier — Knowledge work, any substrate”Claims about cognition and coordination that hold whether the cognizer is a human with paper, a pre-digital office, a modern team, or an AI. They pass the paper-and-pencil test: restate them with “notebook” in place of “file” and they still hold.
Examples from this scaffold:
- Capture → Work → Output — any cognitive system has raw intake, refinement, and stable output regimes. Notebooks have inboxes and archives.
- Temperature gradient — recently-used material deserves easier access than rarely-used. Applies to desks, filing cabinets, and filesystems.
- Attention is finite — the basis of progressive disclosure.
- Convention compresses past decisions — convention principle derives here; language itself is the proof.
- Meta / self-reference is a property of any sufficiently expressive system — meta-self-reference principle.
- The five strata themselves — a claim about conventions in general, not about LLMs.
See the philosophical lineage table — zero-tier claims usually have deep prior thinkers (Drucker, Simon, Wittgenstein, Luhmann).
Tier 1 — Knowledge work with digital technology
Section titled “Tier 1 — Knowledge work with digital technology”Claims that derive from zero-tier but assume addressable digital storage, networks, programmable tools. Would still apply to a pre-AI office computer setup. Wouldn’t apply to a paper office.
Examples from this scaffold:
- Single canonical addressability — principle. Zero-tier ancestor: “ambiguous reference breaks coordination.” Tier-1 version: “filesystems should have one canonical path per resource.”
- Progressive disclosure — zero-tier claim (attention finite) becomes tier-1 once you’re applying it to hyperlinked/hierarchical digital docs.
- Four channels of context — partially tier-1 (notification, query, context-reload patterns) and partially tier-2 (LLM-specific context windows).
Tier 2 — Agentic / LLM-assuming
Section titled “Tier 2 — Agentic / LLM-assuming”Claims that depend on current AI tooling — context windows, agent architectures, skill/tool loading, model invocation.
Examples from this scaffold:
- Skills vs Agents — directly tied to how LLM tooling separates passive expertise from active execution. Zero-tier ancestor: Ryle’s “knowing-how vs knowing-that.”
- Context engineering — assumes LLMs with finite context windows.
- Most of
02-stack/— opinionated patterns for Claude Code + agent tooling.
The paper-and-pencil test
Section titled “The paper-and-pencil test”When writing or auditing a principle, rewrite it substituting:
- “file” → “notebook page”
- “filesystem” → “filing cabinet”
- “agent” → “assistant”
- “context window” → “working-memory capacity”
- “LLM” → “reasoning agent”
If the rewritten version still holds, the principle is zero-tier. If it only holds in the digital version, it’s tier 1 or 2.
This isn’t about taking zero-tier more seriously than tier-2. It’s about knowing what’s stable.
Why this matters
Section titled “Why this matters”1. Longevity
Section titled “1. Longevity”Zero-tier content should never need rewriting unless our understanding of cognition itself changes. Tier-2 content will need updating as AI tooling evolves. Tagging the tier makes “how often does this rot?” visible.
2. Forkability
Section titled “2. Forkability” Someone building paper-based knowledge work → inherits zero-tier only Someone building digital-but-not-AI system → inherits zero + tier 1 Someone building agentic system → inherits zero + tier 1 + tier 2A fork-user who doesn’t use AI still gets value from zero-tier + tier-1 principles if they’re flagged. Without the flag, they have to filter.
3. Epistemic honesty
Section titled “3. Epistemic honesty”Without this distinction, every principle looks equally solid. Tagging tier-of-abstraction lets AI agents and readers weight advice by stability.
Relation to the five strata
Section titled “Relation to the five strata”These are orthogonal axes. Strata measure repeatability of conventions (how much of a convention ports verbatim). Tiers of abstraction measure scope of applicability (what substrate a claim assumes).
| Stratum = how much ports | Tier = what it assumes | |
|---|---|---|
| Same page can be | S1 (philosophy), S2 (pattern), …, S5 (instance) | T0 (zero), T1, T2 |
| Question answered | ”If I copy this, how much changes?" | "Where does this claim apply?” |
| A zero-tier S2 pattern | Pattern-shaped; substrate-agnostic | e.g., “Capture → Work → Output” framework shape |
| A tier-2 S5 instance | Live instance content; LLM-assuming | e.g., “my current Claude Code agent config” |
See the five-strata principle for the stratum axis and its imperfection caveat — both axes are lenses, not taxonomies.
Imperfection caveat
Section titled “Imperfection caveat”Like the five strata, this three-tier classification is a lens, not a taxonomy. Many principles straddle:
- Progressive disclosure has a zero-tier core (finite attention) AND a tier-1 application (hyperlinked docs) AND a tier-2 application (LLM context management). Tag it by the dominant framing in the text.
- SEACOW is structurally zero-tier (any system has entities, activities, medium) but most worked examples are tier-1/2.
- The abstraction hierarchy itself — tiers 0/1/2 — is borrowed from computer science (OSI stack, abstraction ladders) which is tier-1 at minimum. Even naming this framework commits us to a tier-1+ reading.
When in doubt, mark higher (more abstract / zero-tier) — it’s easier to add tech-specificity later than to retract it.
How to use this in practice
Section titled “How to use this in practice”When writing a principle:
- Apply the paper-and-pencil test.
- Tag
tier_of_abstraction: 0 | 1 | 2in frontmatter. - Name the zero-tier ancestor in the body if the principle isn’t zero-tier itself.
- Link to PHILOSOPHICAL-ALIGNMENT for the thinker(s) behind the claim.
When auditing:
- Grep frontmatter for
tier_of_abstraction: 0— these are your most stable claims. - Grep for
tier_of_abstraction: 2— these are your rot-prone claims; plan for rewrites.
When forking:
- Inheriting zero-tier: safe anywhere. No filter needed.
- Inheriting tier-1: assumes your fork uses digital tools.
- Inheriting tier-2: assumes your fork uses AI tooling similar to Claude Code / OpenCode.
See also
Section titled “See also”- Philosophical Alignment — the prior thinkers behind each claim
- Five Strata of Repeatability — the orthogonal axis (how much ports)
- Philosophy — the scaffold’s claims about knowledge work overall
- Meta / Self-Reference — this page itself is a worked example of a meta-principle