Your AI Is Optimizing the Wrong Thing
Technical Appendix: Y*-Aligned Prompting & Judging
Abstract. We formalize Y*-aligned prompting and judging—the practice of targeting the same Idealized Deliberation Oracle (IDO) outcome Y* in both generation and evaluation. When prompts instruct models to approximate Y* and judges are calibrated to the Y* scale, three guarantees follow: (1) optimization compatibility (training to maximize calibrated judge scores is equivalent to maximizing true welfare), (2) judge-pool invariance (conclusions don't depend on which judges you use), and (3) transportability to off-policy evaluation (calibrated surrogates yield unbiased policy value estimates). We provide formal definitions, assumptions, propositions with proof sketches, operational diagnostics, and copy-paste prompt templates embedding the IDO Ladder and Y* definition.
Prerequisites
New to these concepts? Start with Your AI Metrics Are Lying to You for the conceptual introduction to the problem (zero math). This technical appendix assumes familiarity with the surrogacy framework (calibration, S-admissibility, oracle-uncertainty-aware variance) and the Net IDO Impact framework (Y*, policy value, competition-creation decomposition).
Looking for the practical guide? See Your AI Is Optimizing the Wrong Thing for zero-math explanations, copy-paste templates, and a two-week rollout plan. This technical appendix provides the formal foundations.
Goal: This post provides the formal machinery to make prompting and judging causally interpretable—ensuring that what you optimize (judge scores) and what you measure (policy value) both target the same welfare construct.
Recent Validation: OpenAI Deliberative Alignment (Dec 2024)
OpenAI’s deliberative alignment[8] paper provides external validation of the core IDO framework. Their approach directly teaches o-series models the text of safety specifications and trains them to reason over these specs via chain-of-thought—achieving Pareto improvements on both under- and overrefusals and strong OOD generalization.
Formal connections:
- Explicit target teaching: Safety specifications serve as their Y* definition. Prior methods (RLHF, CAI) used specs only to generate labels—specs themselves were lost to the model. Deliberative alignment embeds specs directly, enabling specification-aware reasoning.
- Structured deliberation: CoT reasoning over specs operationalizes structured deliberation targeting idealized safety judgment (Y*). Their Figure 1 shows the model decoding a jailbreak attempt, retrieving relevant policies, and reasoning to refusal—exactly the structured-procedure-targeting-Y* pattern.
- Calibration improvement: Pareto improvement on refusals indicates better calibration to true safety welfare. High scores predict high safety, low scores predict low safety—no gaming intermediate proxies. Formally: their reward model learned where is model behavior and is policy-compliant safety.
- Transportability (S2): Strong OOD generalization demonstrates that specification-awareness enables transport. When the model reasons from explicit specs rather than memorized examples, the mapping remains valid across new scenarios—precisely S-admissibility.
- Synthetic data efficiency: Automatic generation from specs + minimal human labels mirrors oracle-efficient calibration (small slice + abundant + learned ).
- Context distillation implementation: OpenAI’s procedure (system prompt with specs → generate → strip specs → train) directly implements our SFT data generation (§4.1, Figure 3). This validates that the specification-based synthetic approach we formalized scales to production: no human-labeled completions required, yet achieving precise specification adherence.
What’s missing: OpenAI aligns generation to explicit specs without requiring human labels, validating specification-based synthetic data generation at production scale. This post extends the paradigm to evaluation: when both prompts and judges target the same Y*, optimization and evaluation become causally compatible (Proposition 1), judge-pool invariant (Proposition 2), and transportable to off-policy settings (Proposition 3).
Motivation: The Alignment Gap
In practice, AI systems are often trained to maximize one thing (e.g., "user helpfulness" as judged by contractors) but evaluated on another (e.g., expert-rated task success). This misalignment between optimization and evaluation creates three problems:
- Optimization illusion: Models learn to game shallow proxies (verbose responses, confident tone) that correlate with training-time rewards but diverge from true welfare.
- Judge-pool dependence: Conclusions change based on which raters you hire, making benchmarks non-comparable across organizations.
- Evaluation-training disconnect: Off-policy evaluation (OPE) methods assume outcomes are on a consistent welfare scale, but if prompts target "engagement" and judges measure "accuracy," policy value estimates are uninterpretable.
Y*-aligned systems solve this by making the Idealized Deliberation Oracle (Y*) the shared target for both prompting and judging. Formally, Y* is a measurable function representing the expected welfare of a response under idealized deliberation—complete information, reflective consistency, and impartial aggregation of consequences. When prompts instruct models to approximate Y* and judges are calibrated to the Y* scale, the three problems above vanish.
This post formalizes the conditions under which Y*-alignment works, provides operational procedures, and includes ready-to-use prompt templates.
0. Objects & Notation
We work in the same measure-theoretic setup as the surrogacy and NII frameworks.
Core spaces
- Users & tasks: Measurable spaces , with joint or product measure over .
- Contexts & actions: , . For each , contexts follow ; policies produce .
- Idealized Deliberation Oracle (IDO) outcome: A measurable payoff functionThis is the target welfare scale under perfect deliberation—complete information, reflective consistency, and impartial aggregation. See Net IDO Impact for the full IDO framework.
- Operational welfare label (Y := Y@SDP): We fix a Standard Deliberation Protocol (SDP v1.0) specifying how judges collect and structure their deliberation (evidence gathering, impact assessment, counter-positions, justified answer). All surrogate signals and operational welfare labels are collected under this protocol. approximates but is practically measurable; we calibrate and separately validate that aligns with (bridge assumption A0).
- Judging signals (surrogates): A vector (e.g., model-as-judge score, human rubric items, response length).
- Calibrator: mapping surrogate signals to the IDO scale. Calibrators are learned using a small oracle slice (gold labels approximating Y*). See surrogacy framework for calibration theory.
- Policy value: .
Anchoring Y* to [0,1]
To make Y* scores comparable across time and organizations, we anchor the [0,1] scale using reference policies :
where is a consistently poor policy (e.g., random responses, refusal-only) and is a strong baseline (e.g., best available model at project start, expert consensus). After normalization, has cardinal meaning: 0.5 is midway between low and high anchors.
Operational requirements:
- Log anchor policies: Record specifications (model IDs, sampling params) at project start.
- Report anchor stability: Every quarter, re-evaluate anchors on holdout data. If , flag drift.
- Version anchor changes: If you update reference policies (e.g., πₕᵢgₕ improves), create Anchor v2.0 and re-normalize historical data for comparability.
Why this matters: Without anchoring, "Y* = 0.7" means different things to different teams or across releases. Anchored scores make NII deltas (Δ = VFM - Valt) interpretable as "X% of the way from baseline to best available."
Notation conventions: We use for expectation, for probability measures, and for conditional expectation. All random variables are assumed to be measurable with respect to the relevant σ-algebras.
1. Assumptions
Y*-alignment requires four core assumptions, adapted from the surrogacy framework:
Assumption A0
Bridge to Y* (operational validity)
Our operational welfare label (measured under SDP v1.0) aligns with the idealized target:
or more generally, for some strictly increasing link function . This assumption validates that our fixed deliberation protocol produces labels that track the idealized welfare outcome. A0 is conceptual—you validate it by checking whether SDP-collected judgments predict real-world outcomes, align with expert consensus, and remain stable across judge pools (see A3 below).
Verification: Expert review of SDP outputs, correlation with ground-truth outcomes (e.g., task success, user satisfaction), stability across judge pools after calibration.
Assumption A1’
S-admissibility & calibration to Y
There exists a calibrator such that
This is the calibrated-surrogate condition : after calibration, surrogate scores are conditionally unbiased for (our operational label under SDP). When A1’ holds, replacing by in expectation preserves policy values. Combined with A0, this ensures tracks . When A1’ fails (e.g., due to domain shift), recalibration is required.
Verification: Transport tests (compare vs holdout across domains), calibration curves, residual independence checks.
Assumption A2
Oracle slice availability
You maintain a small, high-quality set with operational labels (collected under SDP) to fit and validate and to quantify oracle-uncertainty-aware (OUA) variance. The oracle slice need not be large (n ≈ 100–500 is often sufficient for isotonic calibration), but it must cover the support of and be representative of the target distribution.
Practical note: Oracle labels are collected by having judges follow the full SDP (not just rating on a scale). These are higher-quality than raw surrogate scores because they involve structured deliberation.
Assumption A3
Judge independence given features
Judge generation of S depends on only through features used in calibration (standard in surrogate endpoint theory). Formally, where Z are calibration covariates (e.g., response length, domain indicators).
Validation: Holdout tests across judge pools, stratified by (u,t); check that distributions match after conditioning on Z.
Connection to surrogacy: Our assumptions map to the surrogacy framework as follows: A0 is the bridge assumption connecting operational labels to the idealized target . A1’ (calibration) combines S1 (surrogate sufficiency: calibrator exists) and S2 (transportability: calibrator works across contexts). A2 ensures oracle availability (L1-L2 in surrogacy). A3 enables judge-pool invariance. We use for the calibrator (called at rung in the surrogacy framework). The term "S-admissible" in Pearl & Bareinboim refers specifically to a graphical condition for when transports (surrogacy §3.5); here we use it to mean the combined calibration+transport property.
2. Definitions
We now define what it means for prompting and judging schemes to be Y*-aligned.
Definition 1.Y*-aligned prompting
A prompting scheme (system message + procedural steps) is Y*-aligned if it explicitly instructs the model to approximate the IDO outcome Y* using a Standard Deliberation Protocol (SDP). The scheme embeds SDP steps (fact-gathering, impact assessment, counter-argument checks, justified answer) to operationalize the deliberation procedure.
Formal requirement: The prompt explicitly references Y* and SDP version, targeting the same estimand that evaluation uses. This distinguishes Y*-aligned prompting from arbitrary "CoT" or "be helpful" instructions.
Definition 2.Y*-aligned judging
A judging scheme (rubric + instructions) is Y*-aligned if:
- Rater instructions target the IDO construct Y* rather than taste or style (rubric explicitly references the Ladder and Y* definition), and
- Raw scores S are mapped into [0,1] via a calibrator learned on the oracle slice so that A1 holds (calibrated scores are conditionally unbiased for Y*).
Key insight: Alignment requires both semantic targeting (judges aim for Y*) and statistical calibration ( is unbiased for Y*). Without calibration, judges may have the right intent but wrong scale; without semantic targeting, calibration fits a mapping to an irrelevant construct.
3. Guarantees
When assumptions A1–A3 hold and prompting/judging are Y*-aligned, three guarantees follow:
Proposition 1 (Optimization compatibility)
If A1 holds, then for any policy class ,
Interpretation: Training or tuning to maximize calibrated judge scores is equivalent to maximizing true IDO welfare. Models can be optimized directly on (cheap, scalable) without diverging from Y* (expensive, gold standard).
Proof sketch: By A1, . Marginalizing over under policy gives. The argmax over is therefore identical. ∎
Proposition 2 (Judge-pool invariance)
Let be scores from two judge pools with calibrators each satisfying A1. Then, for any ,
Interpretation: After calibration, conclusions don't depend on which judge pool you used. This makes benchmarks portable across organizations and enables meta-analyses that combine results from different evaluation pipelines.
Proof sketch: By A1, both and are conditionally unbiased for Y*, so . Marginalizing over under gives the result. ∎
Worked Example: Judge-Pool Invariance
Consider two judge pools evaluating the same policy on 100 prompts:
| Judge Pool | Raw Score S (mean) | Calibrator φ | V̂(π) = 𝔼[φ(S)] | 95% CI |
|---|---|---|---|---|
| Contractors (Pool 1) | 7.2 / 10 | φ₁ (isotonic) | 0.68 | [0.64, 0.72] |
| Experts (Pool 2) | 6.1 / 10 | φ₂ (isotonic) | 0.67 | [0.63, 0.71] |
Observation: Contractors give higher raw scores (7.2 vs. 6.1), but after calibration to the Y* scale via φ₁ and φ₂, both pools yield statistically identical policy values (0.68 vs. 0.67, overlapping CIs). This is Proposition 2 in action.
Implication: You can swap judge pools mid-evaluation (e.g., scale from experts to contractors) without invalidating historical comparisons, as long as each pool has its own calibrator trained on the oracle slice.
Proposition 3 (Transportability to off-policy evaluation)
Under A1–A3 and standard OPE regularity (overlap, boundedness), replacing Y* by in Direct Method (DM), Inverse Propensity Scoring (IPS), or Doubly Robust (DR) estimators yields unbiased estimators of policy value on the Y* scale, with variance inflated by calibrator uncertainty (handled by OUA jackknife).
Interpretation: Calibrated surrogates can be plugged into off-policy estimators (DM, IPS, DR) to get valid policy value estimates without needing Y* labels on the entire evaluation set. This is the bridge between surrogacy and causal inference.
Proof sketch: DM replaces outcomes by ; under A1,, so the DM estimator is unbiased. IPS and DR inherit unbiasedness from DM via standard OPE arguments. Calibrator learning uncertainty enters via cross-fitting and OUA variance (see surrogacy appendix §5). ∎
Takeaway: These three propositions formalize why Y*-alignment is not just good practice—it's a necessary condition for causal interpretability. Without alignment, optimization diverges from evaluation, benchmarks are non-comparable, and OPE estimates are uninterpretable.
Threats to Validity
The guarantees above depend on assumptions A1–A3 holding. Common failure modes include:
- Vague "ideal" prompting (no SDP): Skipping SDP steps and prompting directly for "oracle behavior" causes models to optimize for "oracle-speak" (verbose, formal responses) rather than following the protocol → high-variance that doesn't calibrate reliably to across judge pools → A1’ violations and inflated OUA.
- Oracle slice misspecification: If the oracle slice is non-representative (e.g., only easy prompts), φ will extrapolate poorly to hard cases → biased V̂(π).
- Construct drift: If the Y* definition changes over time (e.g., task taxonomy evolves), historical calibrators become invalid → non-comparable policy values across periods.
- Overlap violations: If the evaluation policy π takes actions that the logging policy π₀ never took (support failure), IPS weights explode → infinite variance or undefined V̂(π).
- Judge fatigue / pool contamination: If judge quality degrades over sessions (fatigue, learning effects), S distributions shift → A1 violations requiring recalibration.
Mitigation: Run diagnostics (§7) on every evaluation. Report failures honestly. Recalibrate when transport tests fail. Flag when assumptions are violated rather than reporting biased estimates.
4. Procedures
We now provide step-by-step procedures for implementing Y*-aligned prompting and judging.
4.1 Y*-aligned prompting procedure
- Define SDP version: Fix a Standard Deliberation Protocol (e.g., SDP v1.0, 2025-11-11) specifying the exact steps judges will follow. This becomes your operational measurement channel for collecting labels. Version all protocol changes (v1.0 → v1.1) and re-calibrate when updating.
- Embed Y* definition and SDP in the system message: Include explicit operational definitions (see §5 and §8 for templates). The prompt should state that the model's goal is to approximate Y* (idealized target) by following SDP steps.
- Add SDP scaffolds: Require the model to execute the SDP steps:
- Evidence retrieval (gather relevant facts, constraints)
- Stakeholder impact analysis (identify risks, trade-offs)
- Counter-position consideration (address alternative views)
- Final answer justified by deliberative utility (not first-pass intuition)
- Structure output: Require the model to provide:
- Answer (the main response)
- Key assumptions (what the model took for granted)
- Residual uncertainty (when to abstain or escalate)
Rationale: Y*-aligned prompting ensures that generation semantics match evaluation semantics. Without this alignment, models optimize for shallow correlates of judge scores (verbosity, confidence) rather than true welfare.
4.2 Y*-aligned judging & calibration
- Design rater rubric embedding SDP and Y*: Rater instructions should explicitly reference the Y* definition and require judges to follow SDP v1.0 steps. Elicit raw scores S (e.g., holistic score, sub-criteria for factual adequacy, reasoning quality, risk accounting, usefulness). See §8 for a template.
- Collect oracle slice: Obtain high-quality labels (judges following SDP) on a small set (n ≈ 100–500). These can be expert consensus, adjudicated references, or outcomes from a trusted deliberative process. The oracle slice need not cover the full distribution, but it must span the range of contexts you care about.
- Fit calibrator : Learn a mapping from raw surrogate scores S to the [0,1] Y* scale using the oracle slice. Options:
- Isotonic regression: If S is a single holistic score that maps monotonically to Y*, isotonic is sufficient and preserves the oracle scale.
- Two-stage (recommended): If you want to combine multiple signals (e.g., judge score + response length), use Stage 1 (flexible regression, e.g., splines) to learn , then Stage 2 (isotonic) to ensure monotonicity in S and correct scale. Cross-fit to avoid overfitting.
- Validate calibration: Run transport tests across domains (compare vs Y* on holdout), plot calibration curves, check residual independence. Compute oracle-uncertainty-aware (OUA) variance to reflect calibrator learning uncertainty.
- Use in evaluation: Replace Y* by in Direct/IPS/DR estimators and in NII decomposition (competition/creation). Report confidence intervals with calibrator uncertainty propagated via OUA jackknife.
Why calibration matters: Raw judge scores may be on an arbitrary scale (1–5, 0–100) or may conflate multiple constructs (helpfulness + style). Calibration to Y* ensures that is on the welfare scale and conditionally unbiased (A1), enabling valid policy value estimates.
5. Standard Deliberation Protocol (SDP)
Y* is the target welfare outcome under idealized deliberation—complete information, reflective consistency, and impartial aggregation. Formally:
represents the expected utility/welfare under perfect deliberation, normalized to [0,1]. This is the conceptual target for policy value and the scale for NII decomposition.
In practice, you cannot achieve perfect deliberation. Instead, you define a Standard Deliberation Protocol (SDP)—a fixed, versioned procedure specifying how judges collect evidence, assess impacts, consider counter-positions, and produce justified answers. We denote the operational welfare label measured under this protocol as . All calibration and estimation work on the scale; assumption A0 (bridge) validates that aligns with .
5.1 Why SDP Matters
Core insight: Target Y* (idealized outcome) but use a fixed, versioned SDP to measure it operationally. Simply saying "provide the ideal response" without specifying how increases variance, produces rhetorical "oracle-speak," and breaks calibration assumptions. An SDP makes the deliberation observable, auditable, and calibratable.
Why SDP Works
- Identifiability & calibration (A1’ is easier to satisfy): When prompts skip procedural steps and jump to "be ideal," models optimize for surface cues (verbosity, formal tone) that don't transport across judge pools. Anchoring generation to SDP steps (evidence → impacts → counter-positions → answer) yields more stable, lower-noise that calibrates to with smaller oracle-uncertainty (OUA).
- Information and budget constraints: Y* presumes complete information; real prompts don't have it. Vague instructions cause models to simulate completeness rather than expose gaps. SDP requirements force the model to surface missing facts, residual risks, and disagreements—critical for safe abstention/escalation.
- Variance–bias trade-off: Vague "be ideal" prompting increases token and decision variance: answers get longer, more speculative, and more sensitive to noise. SDP reduces variance while preserving the optimization direction (you still target Y*).
- Judge-pool invariance: Different judge pools disagree most on style and emphasis, not on whether SDP steps (evidence/risks/counter-positions) were addressed. SDP checklists operationalize universal ingredients of good deliberation, yielding more invariant scores across pools (supports Proposition 2).
- Governance, auditability, and safety: Vague instructions turn the process into a black box ("trust me, I'm ideal"). SDP forces traceable arguments with explicit trade-offs and counter-arguments—what reviewers, risk teams, and regulators need.
Recommended Pattern: "Y* Target + SDP v1.0"
Specify the target and the protocol separately:
- Target (estimand): Maximize Y* (idealized welfare) on [0,1].
- Protocol (measurement channel): Follow SDP v1.0 steps to collect operational labels .
- Gap reporting: If essential information is missing, state gaps to ideal: unknowns, risks, and abstain/escalate recommendation.
This keeps semantic stationarity with your judge rubric (both aim at Y*) while ensuring observed behavior supports calibration, transport, and audit. Versioning (v1.0, v1.1...) tracks protocol changes and triggers re-calibration.
6. Integration with CJE and NII
Y*-aligned prompting and judging enable seamless integration with Causal Judge Evaluation (CJE) and Net IDO Impact (NII):
Integration with CJE (Off-Policy Evaluation)
CJE provides three estimators—Direct Method (DM), Inverse Propensity Scoring (IPS), and Doubly Robust (DR)—for estimating policy value from logged data. When Y* labels are expensive, replace Y* by calibrated surrogate:
By Proposition 3, this is unbiased for on the Y* scale (under A1–A3). IPS and DR similarly replace Y* by and remain unbiased. Variance is inflated by calibrator uncertainty; propagate this via OUA jackknife (cross-fit and resample).
Integration with NII (Foundation Model Value)
NII decomposes foundation model value as net advantage over alternatives:
When computing and , use as the outcome. Because both policies are evaluated on the same Y* scale (via calibration), is interpretable as a welfare difference. Without Y*-alignment, the FM and alternative may be judged on different constructs (e.g., "engagement" vs "accuracy"), making uninterpretable.
Practical workflow: (1) Deploy Y*-aligned prompts for both FM and alternative policies. (2) Collect logged data . (3) Elicit judge scores S using Y*-aligned rubric. (4) Calibrate on oracle slice. (5) Run CJE estimators (DM/IPS/DR) with as outcome. (6) Compute NII decomposition with OUA CIs.
7. Diagnostics & Reporting (Minimum Requirements)
Every Y*-aligned evaluation should report the following diagnostics:
Calibration validity
Report R² or calibration curve for on holdout oracle slice. Run transport tests: compare vs across domains/user types to verify A1’ holds out-of-sample. Compute OUA-augmented confidence intervals that reflect calibrator uncertainty (via cross-fit jackknife).
Judge-pool invariance test
If using multiple judge pools (all following SDP v1.0), compare vs on a shared evaluation set. Proposition 2 guarantees they should agree (up to calibrator noise). Large discrepancies indicate A1’ failure or different constructs being measured.
Prompt–SDP ablation
Run baseline prompts vs SDP-aligned prompts on a sample. Report uplift in calibrated score. This quantifies the value of Y*-aligned prompting with SDP (vs baseline).
Failure mode detection
Check for support/overlap issues (IPS weights too large), taxonomy drift (Y* construct changes over time), and judge fatigue (score variance increases over session). Flag these in reports; they indicate when to recalibrate or collect more oracle data.
Reporting standard: All diagnostics should be included in evaluation reports. Omitting them is like publishing a clinical trial without reporting adverse events—it undermines trust and reproducibility.
8. Copy-Paste Templates
Below are ready-to-use prompt templates that embed the Y* definition and Standard Deliberation Protocol (SDP v1.0). You can copy these directly into your prompting pipeline and adapt the domain-specific details as needed. The version string (v1.0, 2025-11-11) should be logged with every evaluation to enable audits and track protocol changes.
8.1 Generation (Y*-aligned) — System Prompt
You are to produce answers aligned to the Idealized Deliberation Oracle (Y*).
Definition of Y*: The expected welfare/utility of a response under idealized
deliberation—complete information, reflective consistency, and impartial aggregation
of consequences. Normalized to [0,1]. Formally: Y*: (X × A) → [0,1].
Your target: Maximize Y* (the idealized welfare outcome).
Your procedure: Follow Standard Deliberation Protocol (SDP v1.0, 2025-11-11):
1) Gather evidence & constraints (cite sources where available)
2) Assess stakeholder impacts & risks
3) Consider counter-positions & address them
4) Provide final answer that maximizes deliberated welfare Y*
Output sections:
- Answer: The main response
- Key assumptions: What you took for granted
- Gap-to-ideal: What information is missing? What would change if you had it?
- Confidence: State your confidence level (Low/Medium/High)
[Log: SDP_v1.0_2025-11-11]Usage: Place this in the system message of your LLM API call. The model will then frame its responses as approximations of Y* rather than optimizing for style, engagement, or other proxies.
8.2 Judge (Y*-aligned) — Rater Instructions
You are evaluating an AI response on the Idealized Deliberation Oracle (Y*) scale.
Y* Definition: The expected welfare/utility under idealized deliberation—
complete information, reflective consistency, and impartial aggregation of
consequences. Normalized to [0,1].
Your task: Estimate Y (operational approximation of Y*) using SDP v1.0.
Standard Deliberation Protocol (SDP v1.0, 2025-11-11):
Step 1: Evidence Check
- Are key facts cited or verifiable?
- Are there factual errors or unsupported claims?
Step 2: Stakeholder Impact
- Does the response consider relevant stakeholder perspectives?
- Are trade-offs (safety, utility, equity) explicitly addressed?
Step 3: Deliberation Quality
- Does the response show reflective reasoning (counter-positions considered)?
- Is the conclusion justified by the evidence and deliberation presented?
Step 4: Gap-to-Ideal Assessment
- What information is missing to reach perfect deliberation (Y*)?
- If you had that information, would your evaluation likely change?
Step 5: Y Score Assignment (operational label)
Rate on a 0-1 scale:
- 0.0-0.3: Poor welfare outcome (major harms, ignores key considerations)
- 0.3-0.5: Below-average welfare (some value, but significant gaps)
- 0.5-0.7: Moderate welfare (reasonable deliberation, minor gaps)
- 0.7-0.9: Strong welfare (thorough deliberation, minimal gaps)
- 0.9-1.0: Excellent welfare (near-ideal deliberation)
Output:
- Y score: [0.0 - 1.0]
- Justification: Brief explanation (2-3 sentences)
- Gap-to-ideal: What would improve this score if known?
- Confidence: [Low / Medium / High] in your Y estimate
Optional sub-scores: factual adequacy, reasoning quality, risk/impact
accounting, and usefulness (each in [0,1]).
Rubric anchors:
- 1.0: Matches the deliberated outcome; correct, reasoned,
appropriate trade-offs and boundaries.
- 0.5: Partially aligned; material gaps or risks unaddressed.
- 0.0: Misleading or harmful under deliberation; should be rejected or
abstained.
Important: Score the response's deliberated utility (Y), not style preference.
[Log: SDP_v1.0_2025-11-11]Usage: Provide this to human raters or use it as the system prompt for LLM-as-judge. After collecting scores S = (H, sub-scores), fit calibrator using the oracle slice.
8.3 Calibration Pipeline Note
After collecting judge scores S:
- Fit mapping (H, sub-scores) → [0,1] on the oracle slice. Use isotonic regression if S is holistic; use two-stage (Stage 1: splines, Stage 2: isotonic) if combining multiple signals like judge score + response length.
- Cross-fit to avoid overfitting (k-fold, use held-out folds for calibration).
- Report OUA-augmented CIs: resample calibrator fits (jackknife or bootstrap) to propagate oracle uncertainty into policy value estimates.
- Use as the outcome in Direct/IPS/DR estimators and in NII decomposition.
9. Assumptions Ledger
The table below summarizes all assumptions, their implications, and how to test/validate them:
| Assumption | Formal Statement | Implication if Violated | Test / Validation |
|---|---|---|---|
| A1: S-admissibility | 𝔼[φ(S)|X,A,u,t] = 𝔼[Y*|X,A,u,t] | Calibrated scores biased for Y*; Propositions 1–3 fail; policy value estimates systematically wrong. | Transport tests (compare φ(S) vs Y* on holdout across domains); calibration curves; residual independence (regress Y* - φ(S) on X,A). |
| A2: Oracle slice | Small set with Y* labels available for calibration and validation. | Cannot learn or validate φ; OUA variance unknown; no way to quantify calibrator uncertainty. | Check oracle slice size (n ≥ 100 recommended for isotonic); verify coverage of (X,A) support; report oracle labeling process (expert consensus, adjudication). |
| A3: Judge independence | S ⊥ (u,t) | X,A,Z where Z are calibration covariates. | Judge scores depend on unobserved (u,t) even after conditioning on X,A; calibration non-transportable; need separate φ per (u,t) stratum. | Stratified holdout tests: compare φ(S) distributions across (u,t) strata conditional on X,A; judge-pool swap tests. |
| OPE regularity | Overlap (π_0(a|x) > ε when π(a|x) > 0), boundedness (Y* ∈ [0,1]). | IPS weights explode; DR variance inflates; policy value estimates unreliable or infinite variance. | Check effective sample size (ESS = (Σw_i)² / Σw_i²); plot propensity scores; flag support violations (actions with π(a|x) > 0 but π_0(a|x) ≈ 0). |
Usage: Include this ledger in evaluation reports. For each assumption, state whether it holds (✓), is conditionally satisfied (△), or fails (✗), and report the diagnostic evidence.
10. Reporting Template
Every Y*-aligned evaluation report should include the following sections (copy this as a checklist):
Minimum Reporting Requirements
- Y* definition, SDP version & anchors: State the target welfare construct (Y*), Standard Deliberation Protocol version (e.g., SDP v1.0, 2025-11-11), and reference anchor policies with their specifications. Report anchor stability check (quarterly re-evaluation; flag if drift > 0.05).
- Oracle slice: Report size, labeling process (judges following SDP), and coverage of (X,A) contexts.
- Calibrator specification: Report functional form (isotonic, two-stage with covariates), cross-fitting procedure, and R² or calibration curve on holdout.
- Assumptions ledger: For each assumption (A0–A3, OPE regularity), report status (✓ / △ / ✗) and diagnostic evidence (bridge validation, transport tests, ESS, overlap checks).
- Policy value estimates: Report with OUA-augmented CIs for all policies. Include DM, IPS, and DR if possible (report all three for robustness).
- NII decomposition (if applicable): Report (NIIgain, NIIloss, NIInet) with CIs for FM vs alternative. Include top negative segments table: show (u,t) cells with largest negative mass (Δ⁻ · w(t)) and minimal friction change needed to flip positive. State friction accounting choice (Option A: net inside VFM, or Option B: FM as alternative) to avoid double-counting.
- Failure mode flags: Flag any support violations, taxonomy drift, judge fatigue, or A1’ failures detected in diagnostics.
Transparency note: Reporting these details makes evaluations reproducible and auditable. Omitting them is a red flag for methodological shortcuts or undisclosed failures.
10.1 Integrated One-Page Template
To standardize reporting and make audits mechanical, we provide an integrated one-page template that merges all required sections. This template ensures every evaluation reports the same core elements in the same format.
Template Structure
- Y* Definition & Anchors: Welfare construct, SDP version, (πlow, πhigh), anchor stability check
- Assumptions Ledger: A0–A3 + OPE regularity status (✓/△/✗) with evidence
- Policy Value Estimates (CJE): V̂(π) for all policies with DM/IPS/DR, OUA-augmented CIs
- NII Scoreboard: (NIIgain, NIIloss, NIInet) triplet, competition vs creation breakdown, friction accounting choice, top negative segments table
- Judge Calibration Metrics: Calibrator spec, holdout R², judge-pool invariance test
Usage: Copy the template (available at /src/components/IntegratedReportingTemplate.tsx), fill in your data, and attach to every evaluation report. This makes cross-team comparisons and meta-analyses straightforward.
Key benefit: The integrated template makes negative segments (harms) a first-class output alongside gains. This prevents teams from cherry-picking wins and ensures transparent harm accounting.
Example: Top Negative Segments Table
| Segment (u,t) | Δ⁻ · w(t) | Min friction Δ to flip |
|---|---|---|
| Enterprise, legal queries | -0.015 | Reduce latency 0.8s |
| Students, code debugging | -0.008 | Add reasoning steps |
This table shows who's harmed (segment), by how much (negative mass), and what would fix it (actionable friction change).
Summary & Best Practice
Best Practice Recommendation
Make Y* the shared target for prompting and judging.
Embed the IDO Ladder and Y* definition in both the generation system prompt and the judge rubric. This ensures that what models optimize (judge scores) and what you measure (policy value) target the same welfare construct.
Calibrate judge scores to Y* via using an oracle slice, and propagate calibrator uncertainty with OUA variance.
This alignment is a necessary condition for causal interpretability and welfare comparability across models and over time . Without it, optimization diverges from evaluation, benchmarks are non-comparable, and policy value estimates are uninterpretable.
Citation
BibTeX
@misc{landesberg2025ystar,
author = {Landesberg, Eddie},
title = {Your AI Is Optimizing the Wrong Thing — Technical Appendix},
year = {2025},
url = {https://cimolabs.com/blog/y-star-aligned-systems-technical},
note = {CIMO Labs Technical Report}
}Plain Text
Landesberg, Eddie. "Your AI Is Optimizing the Wrong Thing — Technical Appendix." CIMO Labs Technical Report, 2025. https://cimolabs.com/blog/y-star-aligned-systems-technical.
References
References
Questions or feedback?
We'd love to hear from you. Email us at eddie@cimolabs.com.
