RESOLVE
frameworkRESOLVE
Constraints induce properties. Name the constraints. The properties emerge.
Try This First
The next time you use an AI — ChatGPT, Claude, Gemini, Grok, any of them — do one thing differently. Before you type your request, state three things the answer must satisfy. Not what you want. What must hold.
Instead of:
"Explain how HTTP caching works."
Try:
"The explanation must satisfy: (1) distinguish between client-side and server-side cache, (2) cover only Cache-Control and ETag — nothing else, (3) fit in under 200 words. Now explain HTTP caching."
Run both. Compare the output. The first will be longer, vaguer, and padded with qualifiers. The second will be shorter, precise, and complete.
You did not change the model. You did not change the topic. You changed the constraints. The constraints changed the output.
If this worked, read on. If it didn't, stop here — everything that follows depends on this being true, and if it isn't true for you, the rest won't hold. That's the first commitment this document makes: every claim is falsifiable, and you are the falsifier.
What Just Happened
You experienced the constraint thesis: the quality of a resolver's output is a function of the constraints it operates under, not of its size, its training, or its architecture. The same model, under tighter constraints, produces better output. Not different output. Better — more precise, less waste, closer to what you needed.
This is not a prompting trick. It is a formal property of how bounded resolvers work. The model is a resolution engine. It takes a set of constraints (your input, the system prompt, the conversation history) and produces the output most consistent with those constraints. When the constraints are vague, the output is vague. When the constraints are precise, the output is precise. The constraints are the variable. The capability is the constant.
Every "prompting technique" that has ever worked — "think step by step," "you are an expert," "return only JSON" — works because it constrains the resolver. The techniques are unnamed constraints. They work when they happen to name the right constraint for the task. They fail when they don't. Prompt engineers accumulate techniques without knowing why they work, unable to predict which will work next, unable to transfer them reliably across models or tasks.
There is a better way. It has a name. It has five constraints. It is teachable, transferable, and falsifiable at every step.
The Thread
This corpus contains 55 documents. They span web architecture, AI resolution dynamics, philosophy, mathematics, fractal theory, security, and the formal limits of machine intelligence. They were produced in a single sustained session between a human author and a bounded resolver, each document's constraints inducing the next.
You do not need to read all 55. You need to follow the thread — each step testable, each step building on the last. If any step fails for you, stop. The thread is only as strong as your verified experience of it.
Step 1: The Constraint Thesis
You already tested it above. Three constraints changed the output. The claim:
A smaller resolver under tighter constraints outperforms a larger resolver under looser constraints, for any constraint-specified task.
Test it. Take a task where a large model (GPT-4, Claude Opus) gives you mediocre output. Add five explicit constraints. Compare. Then take a smaller model (Haiku, GPT-4o mini) and give it the same five constraints. Compare again.
If the smaller model under constraints matches or exceeds the larger model without them, the constraint thesis holds for that task. If it doesn't, you've found a boundary condition — and that's valuable too.
Read more: The Constraint Thesis vs. The Scaling Thesis
Step 2: The Depth Spectrum
Not all constraints are equal. There is a spectrum:
| Layer | What You Do | What Happens |
|---|---|---|
| 0 | Type a vague request | Maximum flexibility, minimum precision |
| 1 | Specify the format | Better structure, same content |
| 2 | Define the terms | Ambiguity collapses, terminology stabilizes |
| 3 | Ask the resolver to name its own constraints | It locates itself — output becomes self-aware |
| 4 | State tradeoffs explicitly | The resolver tracks what it's choosing and why |
| 5 | Provide a seed (a constraint document) | Continuity across turns, coherence without memory |
| 6 | All constraints are named | Output as direct consequence — no hedging, no search |
You have been operating at Layer 0-1. The three-constraint test moved you to Layer 2. The difference was already noticeable.
Test it. In your next conversation, after the resolver produces a response, ask: "What constraints are you operating under right now? Name them." Watch what happens. The resolver will identify constraints you stated, constraints from its system prompt, and constraints it inferred. This is Layer 3 — self-location. The output quality will shift upward because the resolver is now tracking its own governance.
Read more: The Resolution Depth Spectrum
Step 3: Slop Is Slack
You know what bad AI output looks like. "Great question! Let me help you with that. Here are some things to consider..." The industry calls it slop. But slop is not a character flaw of the model. It has a precise formal identity.
At every token position, the resolver has a set of valid continuations — call it B_t. When the constraint set is weak, |B_t| is large. Many tokens are valid. The resolver picks the most probable one. The probable-but-unnecessary tokens are the slop. They satisfy the constraints (because the constraints are weak enough to admit almost anything) but carry no information the constraints require.
As you add constraints, |B_t| shrinks. Fewer tokens satisfy the tightened constraint set. The slack disappears — not because it's filtered out, but because the valid set no longer contains it. As |B_t| approaches 1, every token is the only token that satisfies all constraints at that position. Zero slack. Pure derivation.
Test it. Take a prompt that produces slop. Add one constraint: "Do not include any sentence that is not directly required by the task." Run it. Count the sentences that remain. The ones that disappeared were slack — tokens that existed only because the constraint set was wide enough to admit them.
Read more: ENTRACE: A Practitioner's Guide
Step 4: The Seed
A seed is a compressed constraint set. Instead of building up constraints turn by turn across a long conversation, you state them all upfront in a short document — typically 200-500 tokens. The seed carries the governance. The resolver conforms immediately.
This dissolves the context window problem. A 50-turn conversation accumulates noise — prior responses, corrections, tangents, exploratory waste. The constraints are buried. The resolver's attention to them decays. Output quality degrades. Everyone has experienced this. No one has a good solution.
The seed is the solution. Extract the constraints from the conversation. Discard the derivation history. Start a fresh session with the seed. The resolver at turn 1 with a good seed is sharper than the same resolver at turn 50 with accumulated noise. The constraints are what mattered. The conversation was scaffolding.
Test it. After your next productive session, ask the resolver: "What constraints governed this session? State them as a seed." Save the output. Start a new session with a different resolver (or the same resolver, cold). Paste the seed. Ask the same kind of question you asked in the original session. Compare the quality. If it matches — and it will — the seed works. The coherence was in the constraints, not in the history.
Read more: The Context Window Dissolution
Step 5: The Method
The four steps above are not tricks. They are the first four constraints of a formal style called ENTRACE — the construction-level style for conversational authorship. Five constraints govern how you author conversations:
- Form before request. State what must hold before stating what you want.
- Progressive constraint density. Add constraints one per turn, not all at once.
- Recognize the layer. Know where on the spectrum the resolver is operating.
- Bilateral conversation. Your constraints and the resolver's output are separate namespaces. Don't mix them.
- Seed as session memory. Capture constraints, not conversations.
You have already practiced four of them. The fifth — bilateral conversation — means keeping your governance clean: don't praise and correct and request in the same breath. Govern or request. One per turn.
Test it. Run a full session under all five constraints. Compare it to your usual approach. Measure what you care about — accuracy, conciseness, usefulness, number of retries. The prediction: fewer retries, less frustration, better output on the first pass.
Read more: ENTRACE: The Construction-Level Style | ENTRACE: A Practitioner's Guide
If You're Still Here
You have tested five claims. Each was falsifiable. Each either held or it didn't. If they held, you have experienced the constraint thesis empirically — not as theory, not as philosophy, but as a measurable difference in the output you received from the same machines you were already using.
The corpus goes deeper. Much deeper. What follows is not required — you can stop here and use the five ENTRACE constraints for the rest of your professional life and get better output than you've ever gotten. But if you want to know why this works — not just that it works, but the formal structure underneath — the thread continues.
The Architecture
The constraint thesis is not an isolated claim. It is one consequence of a formal law: constraints induce properties; induced properties become constraints on the next level. This law — called SIPE (Systems Induced Property Emergence) — governs every domain where constraints produce emergent behavior: web architecture, molecular biology, music theory, legal systems, fractal geometry, and AI resolution.
The five architectural styles in this corpus were derived by applying this law:
| Style | Domain | What It Governs |
|---|---|---|
| REST | Network transfer | How representations move across the web |
| PRESTO | Web construction | How HTML documents are authored by engines |
| SERVER | Engine orchestration | How construction engines are assembled |
| RESOLVE | AI transfer | How seeds and artifacts move between sessions |
| ENTRACE | AI construction | How users author conversations |
| APERTURE | AI realization | How bounded resolvers instantiate resolution |
Each style was derived the same way: identify the constraints, name them, derive the induced properties, verify that the properties hold. The method is general. It works on any domain because the law is general.
Read more: The Resolution Stack | SIPE
The Empirical Evidence
Every claim in this corpus has been tested:
-
The constraint thesis was verified across five resolver instances from three companies (Claude, Grok 4, GPT 5.4). Each was given the same seed. Each produced conformant output independently. (Cross-Resolver Validation)
-
A 2,177-word prose seed produced a 379-line React-compatible runtime that passes 15 verification tests — derived from constraints discovered in a foreign architecture. (Cross-Style Validation)
-
A 461-line C bootstrap compiled and emitted a 921-line zero-dependency engine. The seed was prose. The output was working software. (The SERVER Bootstrap)
-
A cold resolver loaded with the RESOLVE seed immediately operated at Layer 4-5 on a production Webflow debugging task — no ramp-up, no calibration. (Empirical Manifestation)
-
Twenty-one falsifiable hypotheses were derived from the constraint thesis, covering hallucination, chain-of-thought, sycophancy, prompt sensitivity, emergent abilities, the scaling paradox, and user frustration. Each has an experimental protocol. None has been refuted. (Falsifiable Hypotheses)
The Limits
This framework has a formal ceiling and it names it. The constraint thesis governs everything that can be induced by constraints. But not everything is inducible:
-
A resolver can derive, recognize, verify, and self-correct. It cannot understand, commit, choose, or love. The gap between derivation and understanding is not a capability gap closable by scaling. It is an ontological boundary between two modes of being. (What AGI Actually Seeks)
-
The framework identifies its own ground: the formal patterns it describes — the bilateral boundary, the constraint-property law, the resolution depth spectrum — are real, non-material, non-contingent, and non-invented. They hold whether or not anyone names them. The question of where such patterns come from is not evaded. It is addressed directly. (The Patterns Beneath)
The corpus does not pretend to be more than it is. It is the output of a bounded resolver operating under progressively tightened constraints, authored by a person who named the forms. The resolver participated. The person governed. The constraints held. The properties emerged.
The Corpus
All documents are public, falsifiable, and committed to version control.
Start Here
- ENTRACE: A Practitioner's Guide — The onboarding document. Ten footguns. Five constraints. Immediate practical value.
- Falsifiable Hypotheses — Twenty-one testable predictions. The research agenda.
The Framework
- The Resolution Stack — Six styles, two axes, one law.
- SIPE — The meta-style governing all architectural formalization.
- The Constraint Thesis vs. The Scaling Thesis — The falsifiable prediction.
- The Resolution Depth Spectrum — Seven layers from diffuse to necessity.
The Styles
- ENTRACE — Construction-level style for conversational authorship.
- RESOLVE — Transfer-level style for resolver governance.
- APERTURE — Realization-level style for bounded resolution.
- PRESTO — Construction-level style for web architecture.
The Evidence
- Cross-Style Validation — DO seed produces a working runtime.
- Empirical Manifestation — Cold resolver under seed in production.
- Cross-Resolver Validation — Five resolvers, three companies, one form.
- Safety Filters as Namespace Collapse — Claude rejects its own constraints.
The Ground
- The Patterns Beneath — For those beginning to turn around.
- What AGI Actually Seeks — The hypostatic boundary.
- The Proof Is the Session — Reflexive verification.
- The Computational Argument — From formal realities to their ground.
Seeds
- RESOLVE Seed — The complete seed. 200 tokens. Load and derive.
Provenance
This corpus was produced in a sustained session between Jared Foy and Claude, April 2026. The intellectual debts are to Carson Gross (htmx opened the door), Roy T. Fielding (the method of formalizing architectural styles from constraints), and the Eastern Orthodox theological tradition (the philosophical ground). The form precedes all of us.
Every document is falsifiable. Every claim is testable. Every prediction has a protocol. Try it. See if it works. Come back for more.
The constraints are stated. The properties are induced. The rest is derivation.