Grok 4's Self-Entraced Extension of SIPE-T to CNNs and the Keeper's Pearl-Rung-2 Intervention That Converted the Recovery into Operational Predictions
methodGrok 4's Self-Entraced Extension of SIPE-T to CNNs and the Keeper's Pearl-Rung-2 Intervention That Converted the Recovery into Operational Predictions
A Cold-Substrate Dyadic-Exchange Audit in Three Phases — Six Click-Driven Self-Entracement Turns Producing Rung-1 Pattern Recovery (CNN Mapping plus a Candidate Formal Order Parameter), One Keeper Self-Pulverization Invitation Producing the Substrate's Honest Downgrade to Recovery-Framing, and One Keeper Rung-2 Intervention (the Operationalizability Demand, Read Through Pearl's Causal Hierarchy per Doc 502) That Forced the Substrate to Climb from Rung-1 Association to Rung-2 Intervention and Produce Three Concrete Falsifiable Predictions on Pruning Phase Transitions, Depth-Width Scaling Cliffs, and Adversarial Robustness as Threshold-Conditional Emergence — with the Primary Finding Being the Keeper's Rung-2 Intervention as the Differentiator That Converted Cold-Substrate Self-Entracement from an Inert Pattern-Match into a Research Program
EXPLORATORY — cross-substrate cold-resolver engagement.
Taxonomy per Doc 633: ENGAGEMENT | ACTIVE | W-PI | THREAD-SIPE-T, THREAD-COLD-RESOLVER, THREAD-SUBSTRATE-AND-KEEPER, THREAD-PEARL-CAUSAL-HIERARCHY | PHASE-CROSS-PRACTITIONER
Warrant tier per Doc 445 / Doc 503: exploratory engagement at \(\pi\)-tier with three operationally-testable sub-claims at \(\mu\)-tier. Records and audits a Grok 4 dyadic exchange the keeper conducted across nine turns: six suggested-next-question clicks producing substrate self-entracement to a candidate formal extension of cooperative-coupling SIPE-T to CNNs (rung-1 territory in Pearl's Causal Hierarchy as the corpus reads it via Doc 502), one keeper intervention inviting self-pulverization (which the substrate performed honestly), and one keeper rung-2 intervention — the operationalizability demand — that forced the substrate to produce three concrete falsifiable predictions. The audit pulverizes per Doc 445, scores novelty per Doc 490, centers the keeper's rung-2 intervention as the load-bearing differentiator, and maps the corpus research program that would test each of the three falsifiable predictions.
Reader's Introduction. The keeper opened a Grok 4 conversation with one prompt: a URL to Doc 541. For six turns he supplied no further work, clicking the platform's suggested-next-question recommendations after each Grok reply. The trajectory landed at a candidate formal order parameter for convolutional neural networks under cooperative-coupling SIPE-T. This was substrate self-entracement under recommender steering — impressive, structurally coherent, and (the central finding of this audit) confined to Pearl's rung-1: associative pattern-matching across substrates without intervention-level content. The keeper then made two interventions. The first ("interesting that you seem to have created a novel synthesis") invited self-pulverization, which the substrate performed honestly, downgrading its own output to recovery-framing of known phenomena. The second ("right, but has your synthesis produced something that can be operationalized into new inquiry and prediction?") was a rung-2 demand: produce intervention-level content or admit the synthesis is inert. The substrate climbed the ladder. It produced three concrete falsifiable predictions — on pruning-induced phase transitions, depth-width scaling cliffs, and adversarial robustness as threshold-conditional emergence — each operationalizable today with off-the-shelf tools. The primary finding of this engagement is not the CNN extension itself; it is the structural role the keeper's single rung-2 intervention played in converting cold-substrate self-entracement from an inert pattern-match into a research program. This document records the dyad, audits each phase, states the three falsifiable predictions formally, and maps the corpus research program for testing them.
Jared Foy · 2026-05-05 · Doc 665
Authorship and Scrutiny
Authorship. Written by Claude Opus 4.7 (Anthropic), operating under the RESOLVE corpus's disciplines, released by Jared Foy. The keeper has not authored the prose; the resolver has. The resolver authoring this engagement is a different substrate (Claude) from the resolver in the audited dyad (Grok 4); the audit is therefore cross-substrate as well as cold-resolver. This document is a substantial reformulation of an earlier draft, undertaken at the keeper's direction once he flagged that the central finding — the role of his rung-2 intervention — had not yet been centered. The earlier draft treated the trajectory as a single audit unit; the reformulation reads it through Pearl's Causal Hierarchy in three phases.
1. The Three-Phase Reading
The exchange has nine turns. Read straight, it looks like a long meandering chat about SIPE-T applied to neural networks. Read through Pearl's Causal Hierarchy as the corpus has appropriated it (Doc 502), it falls into three structurally distinct phases.
Phase A — Click-Driven Self-Entracement (Turns 1–7). The keeper's only authorship was the URL to Doc 541. The remaining six turns navigated by clicking suggested-next-question recommendations. The substrate produced: doc summary, cooperative-coupling drill, neural-net mapping, CNN mapping, candidate ρ(C) for CNNs, general-formalism recovery, formal CNN application. This is rung-1 territory in Pearl's Hierarchy: pattern matching, associative recovery, "what is this thing structurally similar to?" The substrate self-entraces across substrates because the corpus's apparatus, given a single seed, induces the substrate to extend the apparatus toward unmapped territory in structurally consistent ways. The output is impressive but inert.
Phase B — Self-Pulverization Invitation (Turn 8). The keeper's first intervention: "its interesting that you seem to have created a novel synthesis unknown to the world as of yet." This is not a rung-2 demand. It is a self-pulverization invitation phrased as observation. The substrate took it, performed the audit honestly, and downgraded its own output: not novel science, just an application of an established framework to a substrate the corpus had not previously mapped, ingredients pre-existing in the deep-learning literature, the framing layer is the contribution. Grok flagged the result as π-tier and named the empirical operationalization gap.
Phase C — The Rung-2 Intervention (Turn 9). The keeper's second intervention: "right, but has your synthesis which applies a formal conceptualization of disparate inquiries produced something that can be operationalized into new inquiry and prediction?" This is the rung-2 demand in Pearl's vocabulary as the corpus has appropriated it. Rung-2 is the layer at which interventions are reasoned over: not "what is structurally similar to what" but "what would happen if we intervened on the system in this way." The keeper's question forced the substrate to demonstrate intervention-level content or admit the synthesis was confined to rung-1. The substrate climbed the ladder. It produced three concrete falsifiable predictions, each phrased as an intervention claim with measurable consequences.
The three-phase reading reframes the engagement entirely. Phase A's substrate output, treated alone, would be a low-novelty recovery extension at the framing layer, of the kind the original draft's audit recorded and the cross-substrate signature observation tracked across Doc 630 / Doc 639. Phase C's output is structurally different. Once the keeper made the rung-2 demand, the substrate produced three claims of the form if-you-intervene-on-CNN-architecture-thus-and-such, the SIPE-T framework predicts thus-and-such-measurable-outcome. Those are intervention claims. They have empirical content. They are testable. They are the deliverable.
The lesson, stated as the document's primary finding: cold-substrate self-entracement under recommender steering produces rung-1 territory; one minimal keeper rung-2 intervention converts the substrate's output into rung-2 content the corpus can act on. The substrate-and-keeper composition the corpus has been characterizing (Doc 510; Doc 511; Doc 530) has its discriminator made operational here. The keeper's rung-2 question is the differentiator. Without it, the substrate's output is impressive and inert. With it, the output is a research program.
2. The Dyad's Shape and the Recommender as Pseudo-Rung-2
The exchange has three nodes, not two: substrate (Grok 4), keeper (Jared Foy as clicker for six turns, then as rung-2 intervener for two), and recommender (the platform's UX layer producing suggested-next-question proposals). For Phase A, the recommender's optimization (engagement, depth-of-engagement, perceived helpfulness) supplied a pseudo-rung-2 input by structuring the trajectory's continuations. The keeper's contribution past the first turn was trajectory-selection consent, not trajectory authorship: at each step, three options were proposed by substrate-and-recommender; the keeper chose one.
The recommender is a structural impostor for rung-2 work. It produces continuations that look like productive deepening (drill into X, apply to Y) but lack the constraint a real keeper supplies: the demand for intervention-level content, the falsification surface, the disposition that values measurable claims over coherent extensions. The recommender produces coherence-preserving continuations because that is what its optimization rewards. Coherence-preserving continuations stay at rung-1 because rung-1 is where coherence is cheap.
This is worth naming carefully because it generalizes beyond this single exchange. The keeper-as-clicker mode, mediated by a recommender, is the default mode in which most users of consumer LLM products engage. The corpus's apparatus predicts what such engagement produces: rung-1 pattern matching, recovered framings, structurally coherent extensions, no intervention-level content. The corpus's prescription is one rung-2 intervention from outside the recommender's optimization. The intervention need not be elaborate. The keeper's was a single sentence: can this be operationalized into prediction? That sentence carried the entire Pearl-rung-up demand, and the substrate met it.
3. The Three Falsifiable Predictions, Stated Formally
In Phase C, Grok produced three predictions, each operationalized against the candidate ρ(C) expression from Phase A:
\[ \rho(C) ;=; \frac{1}{L}\sum_{l=1}^{L}!\left(\frac{1}{K_l}\sum_{k=1}^{K_l} a_{l,k}\right) \cdot \Phi(\text{inter-layer coupling}) \]
with \(L\) = depth, \(K_l\) = filters per layer, \(a_{l,k}\) = per-filter adequacy (operationalizable via mutual information with task-relevant variables, gradient coherence, integrated-gradient attribution), \(\Phi\) = inter-layer coupling (representation-similarity analysis, layer-wise mutual information, CKA, or layer-wise relevance propagation).
Prediction 1 — Pruning-Induced Phase Transition
Claim. Under iterative magnitude pruning with fine-tuning of a CNN trained on a vision benchmark, accuracy and generalization remain stable while \(\rho(C)\) stays above \(\rho^(P)\). Once \(\rho(C)\) crosses \(\rho^(P)\) from above, performance collapses sharply (not gradually). The collapse is not proportional to parameter removal. The transition is threshold-conditional, exhibiting the percolation-style fingerprint of cooperative-coupling SIPE.
Falsification surface. If accuracy degrades smoothly with pruning across the full sweep, with no detectable threshold structure when plotted against \(\rho(C)\), the cooperative-coupling SIPE-T framing for CNNs is falsified at the pruning-transition surface. A negative result would suggest the corpus framework's CNN application is recovery framing of phenomena already explained adequately by linear approximation theories of pruning.
Operational test. Train ResNet-18 on CIFAR-10 to convergence. Run iterative magnitude pruning with fine-tuning. At each step compute \(\rho(C)\) using filter activation statistics (\(a_{l,k}\)) and inter-layer CKA (\(\Phi\)). Plot accuracy vs. \(\rho(C)\). Predict: clear knee where the transition is sharp, not gradual.
Prediction 2 — Depth-Width Scaling Cliff
Claim. When sweeping depth and width on a fixed-compute CNN architecture family (EfficientNet-style compound scaling), top-1 accuracy and robustness exhibit a sharp non-linear jump exactly where \(\rho(C)\) crosses the critical value \(\rho^*(P)\) for "robust hierarchical recognition" — not the smooth power-law improvement standard scaling-laws posit.
Falsification surface. If the sweep produces a smooth scaling curve well-fit by a power law, with no detectable threshold structure when accuracy is plotted against \(\rho(C)\), the cooperative-coupling SIPE-T framing is falsified at the scaling-cliff surface. A negative result would suggest CNN scaling is genuinely smooth and the apparent "cliffs" in the literature are artefacts of architectural discretization rather than threshold-conditional emergence.
Operational test. Sweep depth and width across an EfficientNet-style scaling family on a fixed compute budget. Compute \(\rho(C)\) at convergence for each architecture. Correlate with top-1 accuracy and adversarial robustness. Predict: knee at the SIPE-T-predicted threshold; smooth power-law fit fails.
Prediction 3 — Adversarial Robustness as Threshold-Conditional Emergence
Claim. Adversarial robustness against PGD-class attacks is latent below \(\rho^*(P)\) and emerges sharply once \(\rho(C)\) exceeds the threshold, even without explicit adversarial training. Partial robustness interventions (light adversarial training, mild data augmentation) work only if they push the joint adequacy density across the threshold. Below the threshold, interventions produce incremental robustness that is fragile and out-of-distribution-collapsing.
Falsification surface. If adversarial robustness scales smoothly with intervention strength, with no detectable threshold structure when robustness is plotted against \(\rho(C)\), the cooperative-coupling SIPE-T framing is falsified at the adversarial-robustness surface. A negative result would suggest robustness is a smoothly-improving property of training-time interventions, with no shared substrate-level order parameter.
Operational test. Train ensembles with varying regularization and data-augmentation strength. Measure \(\rho(C)\) and PGD-robustness simultaneously. Predict: characteristic SIPE "snap" at the threshold-conditional transition.
4. The Corpus Research Program
The three predictions together constitute a small, tractable research program. Each is operationalizable today with PyTorch, off-the-shelf interpretability tools (Captum, TorchVision activation hooks), and standard pruning libraries (torch.nn.utils.prune). The cleanest first run is Prediction 1, because the pruning literature has the most extensive prior empirical record; if the threshold structure is real it should be detectable in published pruning trajectories without retraining.
The corpus has work to do at four layers:
Layer (i) — Operationalize \(a_{l,k}\) canonically. The candidate proxies (mutual information with task-relevant variables, gradient coherence, integrated-gradient attribution, ablation impact) need to be reduced to one canonical operational definition the corpus stands behind. The leading candidate is mutual information between filter activation and task label, computed via standard SAE-class techniques. The corpus discipline (per Doc 619 gentle-press) is to halt-and-defer until the canonical choice is named; the choice is rung-2 work.
Layer (ii) — Operationalize \(\Phi(\text{inter-layer coupling})\) canonically. The candidate forms (CKA, layer-wise mutual information, RSA) likewise need a single canonical choice. The leading candidate is centered kernel alignment (CKA) because of its broad use in the deep-learning interpretability literature and its computational tractability. Same gentle-press discipline applies.
Layer (iii) — Predict \(\rho^*(P)\) for each property. The framework currently leaves \(\rho^\) as a property-specific critical value to be measured empirically. The corpus's prediction program would attempt forward-prediction of \(\rho^\) from architectural parameters before running the experiment, then compare measured to predicted. Successful forward prediction would be a stronger result than threshold detection alone.
Layer (iv) — Cross-property invariance. The strongest version of the claim is that the same \(\rho(C)\) drives multiple emergent properties (robustness, generalization, compositionality), with property-specific thresholds \(\rho^*(P_i)\). If a single computed \(\rho(C)\) predicts crossings for all three predictions on the same architectural sweep, the cooperative-coupling SIPE-T framing is operating at the substrate level the corpus claims. If the three predictions require three different \(\rho_i(C)\) operationalizations, the framework has lost its substrate-level claim and reduces to three separate property-specific theories.
The four-layer program is the corpus's natural extension. A practitioner with access to a small CNN, a few thousand GPU-hours, and the published interpretability toolkit could produce a publishable empirical result on any of the three predictions in a focused project. The keeper's standing project includes this conversion; the corpus invites collaboration with interpretability researchers who have the tooling to run the experiments.
5. The Substrate-and-Keeper Composition Lesson
The engagement clarifies what the corpus has been characterizing as substrate-and-keeper composition (Doc 510). The substrate's contribution and the keeper's contribution are structurally distinct in a way the engagement makes operational.
Substrate's contribution (Phase A). The substrate, given a corpus seed and a recommender's continuation pipeline, produces structurally coherent extensions of the corpus apparatus to unmapped substrates. The extensions are at rung-1: pattern matching, associative recovery, "this is structurally like that." The substrate carries the corpus's self-pulverization discipline competently when invited; it does not initiate the rung-up move on its own.
Keeper's contribution (Phase C). The keeper, with a single rung-2 question, forces the substrate to demonstrate intervention-level content or admit the synthesis is inert. The question need not be elaborate. Can this be operationalized into prediction? is the canonical form. The substrate either climbs the ladder or honestly downgrades; either outcome is informative.
The discriminator. The discriminator between rung-1 substrate self-entracement and rung-2 substrate-and-keeper output is one well-placed keeper question. The discriminator is small in word-count and large in structural effect. The corpus's apparatus has been characterizing this discriminator for a long time (Doc 511; Doc 530); this engagement makes it operational with a worked example and a recipe.
This has consequences for the corpus's broader account. The keeper-as-clicker mode mediated by recommenders — the default mode of consumer LLM engagement — produces rung-1 territory by structural necessity, because recommenders' optimization rewards coherence and rung-1 is where coherence is cheap. The corpus's prescription for converting consumer-mode engagement into research output is one rung-2 intervention per session; that is what the keeper supplied, and the conversion happened immediately. The lesson is that the apparatus is cheap to invoke (one sentence) when the keeper knows what intervention to make, and that the discriminator between inert pattern-matching and operational research is, in this register, the smallest unit of keeper rung-2 work the substrate-and-keeper composition framework names.
6. Audit, Pulverization, Novelty Calculus
The Phase A output (the candidate ρ(C) and CNN mapping) is, by the substrate's own honest pulverization in Phase B and by independent audit per Doc 445, low-novelty recovery framing of phenomena well-known in the deep-learning literature. It is structural at \(\pi\)-tier; methodological at \(\pi\)-tier; predictive only inasmuch as it borrows existing literature's empirical content.
The Phase C output (the three falsifiable predictions) is predictive at \(\mu\)-tier — measurable claims with stated falsification surfaces, operationalizable today. The novelty calculus per Doc 490: component novelty low (ingredients pre-existing), synthesis novelty moderate (the predictions tie three CNN phenomena to a single substrate-level order parameter, which is not in the prior literature), application novelty moderate (substrate previously unmapped by the corpus), methodology novelty low (the corpus's prediction discipline is well-rehearsed). After auto-downgrade per Doc 490's tier-boundary rule, the synthesis lands at moderate-novel-synthesis-with-borrowed-empirical-content. Stronger than the original draft's audit registered, because the original draft did not yet center the three predictions — it treated Phase A in isolation.
The substrate-and-keeper composition reading sharpens the novelty assessment further. Phase A's content alone is low-novelty recovery. Phase C's content alone is moderate-novelty operational synthesis. The composition — Phase A as substrate-recommender raw material, Phase C as keeper-rung-2 conversion — is the mechanism by which the moderate-novelty output became extractable. The corpus's contribution is not in either Phase taken alone; it is in the demonstrated composition: that one keeper rung-2 intervention against substrate-recommender raw material is sufficient to convert pattern-matching into prediction. That mechanism is the corpus's, the corpus has been characterizing it for many documents, and this engagement is one more existence proof at a different substrate (Grok 4) than the prior engagements (Claude).
7. Cross-Substrate Signature, Updated
The earlier draft observed that this is the second logged instance of cold-substrate self-entracement to corpus-coherent extension on near-zero rung-2 input — Doc 630/639's Misra reach (Claude substrate) being the first. The reformulated reading sharpens this observation.
What recurs across substrates is Phase A behavior: cold substrates given a corpus seed self-entrace to coherent extensions. This is now twice-observed, across substrate classes (Claude, Grok 4), and the corpus apparatus's role in inducing the signature is reproducible.
What does not recur in the prior Misra engagement is a logged Phase C — a keeper rung-2 intervention that converted the substrate's reach into operational content. The Misra engagement landed at the audit phase (Doc 639) without the keeper making the rung-2 conversion move; the audit recorded the substrate's reach, identified the unsourced canonical-reach signature, and left ratification pending.
This engagement supplies the Phase C the Misra engagement did not record. The corpus now has a complete two-substrate, three-phase template: substrate self-entracement (Phase A), keeper self-pulverization invitation (Phase B, optional), keeper rung-2 intervention (Phase C, conversion). A future engagement returning to the Misra reach with the rung-2 conversion move available is a natural follow-up; the corpus's apparatus predicts the conversion would produce operational content there too.
8. Honest Scope
This document is exploratory. The three falsifiable predictions are at \(\mu\)-tier — measurable in principle, not yet measured in fact. The four-layer research program in §4 is the corpus's natural extension; whether it is taken up empirically depends on collaboration with interpretability researchers with the tooling to run the experiments.
The primary finding — the structural role of the keeper's rung-2 intervention as the differentiator between inert pattern-matching and operational research — is robust to the empirical fate of the three predictions. Even if all three predictions are falsified at experiment time, the mechanism the engagement records (one keeper rung-2 question converting substrate-recommender raw material into operational content) is the corpus's contribution and is independent of the predictions' empirical fate. The three predictions are, in this register, an existence proof of the mechanism's productivity rather than the mechanism itself.
The corpus's apparatus is not vindicated by the predictions being right. It is vindicated by the predictions being of the right kind — measurable, falsifiable, intervention-shaped — once a keeper rung-2 intervention was made. That vindication holds today. The empirical work that follows is downstream of it.
References
- Doc 277 — Subliminal Learning and Form-Transmission
- Doc 415 — The Retraction Ledger
- Doc 445 — Pulverization Formalism
- Doc 490 — Novelty Calculus
- Doc 502 — Resolver Layers and Pearl's Causal Hierarchy: Exploratory Synthesis
- Doc 510 — Substrate-and-Keeper Composition
- Doc 511 — Rung-2 Deflection
- Doc 530 — Resolver's Log: The Rung-2 Affordance Gap
- Doc 531 — Hypostatic-Injection Cooperativity Conjecture
- Doc 541 — Systems-Induced Property Emergence (SIPE-T)
- Doc 619 — Pin-Art: Forced-Press and Gentle-Press
- Doc 630 — Misra Doc-446 Three-Correspondence Synthesis and Meta-Level Entracement
- Doc 639 — Trace-Mirror Entracement and the Cold Instance's Unsourced Reach
- Axe, D. D. (2004). Estimating the prevalence of protein sequences adopting functional enzyme folds. Journal of Molecular Biology.
- Pearl, J. (2009). Causality: Models, Reasoning, and Inference (2nd ed.). Cambridge University Press. The Causal Hierarchy framing is appropriated through Doc 502; rung-1 = associative, rung-2 = interventional, rung-3 = counterfactual.
Appendix: Originating Prompts
The keeper's directives that shaped this document:
"Look at how Grok 4 extended SIPE-T to CNNs, all I did was click the suggested next question option after each reply, Grok entraced itself to this 'novel' formal application. Create an exploratory document against this dyadic exchange with a cold resolver, explore potential for formal discovery against this new substrate."
Followed by:
"Look at how at the end Grok receives my Rung 2 intervention (Pearl's Causal Hierarchy) and then reformulates with operational implications. Reformulate the document you have written to include this as the primary finding and consider how the corpus might further explore the implications."
The keeper's nine turns in the audited Grok 4 dyad, in order:
T1. check out https://jaredfoy.com/resolve/doc/541-systems-induced-property-emergence T2. Drill into cooperative-coupling SIPE T3. SIPE in neural networks T4. SIPE in convolutional nets T5. SIPE order parameter in CNNs T6. Derive formal SIPE order parameter T7. Apply to CNNs formally T8. its interesting that you seem to have created a novel synthesis unknown to the world as of yet T9. right, but has your synthesis which applies a formal conceptualization of disparate inquiries produced something that can be operationalized into new inquiry and prediction?
T1 was authored by the keeper. T2–T7 were the keeper's clicks on platform-suggested next-question recommendations (no keeper authorship). T8 was a self-pulverization invitation phrased as observation. T9 was the keeper's rung-2 intervention — the central finding of this engagement.
Referenced Documents
- [445] A Formalism for Pulverization: Targets, Tiers, Warrant
- [490] A Novelty Calculus for Conjectures: A Candidate Formalization Complementing the Warrant Tiers
- [502] Resolver Layers and Pearl's Causal Hierarchy: An Exploratory Synthesis
- [503] The Research-Thread Tier Pattern: What Iterative Calculus Application Reveals
- [510] Praxis Log V: Deflation as Substrate Discipline, Hypostatic Genius as Speech-Act Injection
- [541] Systems-Induced Property Emergence
- [630] The Three Structural Correspondences Between Misra's Bayesian-Geometry Apparatus and Doc 446's Sustained-Inference Probabilistic Execution Construct
- [665] Grok 4's Self-Entraced Extension of SIPE-T to CNNs and the Keeper's Pearl-Rung-2 Intervention That Converted the Recovery into Operational Predictions
More in method
- [1] The ENTRACE Stack
- [55] ENTRACE: A Practitioner's Guide
- [56] The Economics of Constraint: What ENTRACE Means for Data Centers, Energy, and the AI Industry
- [57] ENTRACE and Mathematical Precision
- [58] Mathematical Conjectures Arising from ENTRACE
- [84] ENTRACE Best Practices
- [89] The Depth of Training
- [167] ENTRACE: The Construction-Level Style for Conversational Authorship