Halaman

Senin, 23 Maret 2026

Artikel: Epistemic Signal Ratio

Epistemic Signal Ratio: A Process-Native Measure of Reasoning Quality in Decision-Making


Toward a Causal Theory of Reasoning Accountability in Governance


Accountability‑Based Universal Wisdom and Trust

Cross-Sector Pre-Decision Governance Translator


Final Manuscript · March 2026


---


ABSTRACT


Contemporary governance systems have developed sophisticated mechanisms for financial accountability, procedural compliance, and performance measurement. Yet, a persistent paradox remains: decisions that meet all formal governance standards often fail substantively. This paper argues that existing governance frameworks cannot rigorously identify decision-relevant reasoning under counterfactual criteria—they lack a systematic way to distinguish reasoning that actually influences decisions from reasoning that merely accompanies them. To address this gap, we introduce a formalized and counterfactually grounded class of metrics called process-native epistemic metrics, with the Epistemic Signal Ratio (ESR) as its first instantiation. ESR measures the proportion of documented reasoning that genuinely affects decision outcomes, defining relevance through counterfactual perturbation: an information unit is relevant if its removal changes the ranking of decision alternatives. We provide a reduced-form micro-foundation grounded in diminishing marginal cognitive returns, establish formal theorems on information overload, identifiability under panel aggregation, causal identification under exogenous documentation shocks (LATE), and groupthink amplification. We show that ESR is a partial sufficient statistic for manifest reasoning efficiency conditional on a closed information set, prove its non-equivalence to Shapley value (ordinal vs cardinal invariance), derive testable restrictions (\beta_1 > 0, \beta_2 < 0), and analyze strategic manipulation: adding noise is a dominated strategy (ESR is incentive‑compatible with respect to noise addition), though selection manipulation remains possible (ESR is not fully strategy‑proof). ESR thus belongs to the class of partially robust governance metrics. Extending beyond efficiency, we introduce Latent Signal (S_L) and Epistemic Blindspot (EB) to measure what organizations fail to consider, defining latent signal relative to ex post validation rather than pure speculation (excluding hindsight‑only constructs). The framework offers tiered implementation (Lite, Internal BCEV+, Full BCEV+), empirical test mapping, and adversarial audit protocols. ESR transforms reasoning quality from an unobservable cognitive construct into a statistically estimable and causally testable variable, contributing to measurement theory, decision theory, and governance theory.


Keywords: epistemic governance, pre-decision accountability, counterfactual relevance, bounded rationality, causal identification, epistemic blindspot, groupthink, partially robust metrics


---


1. INTRODUCTION


1.1 The Paradox of Procedural Compliance


Modern governance systems have evolved sophisticated accountability mechanisms: financial audits ensure proper resource use, procedural reviews mandate compliance, and performance evaluations measure outcomes. Yet, a persistent paradox remains: decisions that formally satisfy all three dimensions—financially compliant, procedurally correct, and meeting performance targets—frequently fail substantively. Infrastructure projects approved through rigorous feasibility studies become stranded assets. Social programs meeting output targets fail to achieve intended impact. Strategic investments vetted through due diligence result in catastrophic losses.


This paradox—formal compliance yet substantive failure—has been examined through multiple lenses. Herbert Simon (1947) introduced bounded rationality, showing that decision-makers operate under cognitive constraints. Kahneman and Tversky (1979) identified systematic biases that affect judgment. The behavioral public administration literature (Grimmelikhuijsen et al., 2017) has documented how cognitive biases interact with bureaucratic structures. Yet, despite these insights, governance frameworks remain focused on what was decided and whether procedures were followed, not on how reasoning was structured and whether it actually mattered for the decision.


1.2 The Identification Problem


We argue that existing governance systems face a fundamental identification problem with respect to decision-relevant reasoning. A governance system can identify decision-relevant reasoning if it can distinguish:


· Epistemic signal: reasoning that genuinely influences decision outcomes

· Epistemic noise: reasoning that is present but does not affect outcomes


Without a counterfactual definition of relevance—whether removing an information unit would change the ranking of alternatives—existing approaches cannot rigorously identify which reasoning matters. Subjective importance ratings, readability metrics, and information volume measures all fail to establish whether documented reasoning actually influenced the final decision. This is not to say that current systems are completely blind; rather, they cannot rigorously identify decision-relevant reasoning under counterfactual criteria.


1.3 A Formalized and Counterfactually Grounded Class of Metrics


This paper introduces a formalized and counterfactually grounded class of governance metrics: process-native epistemic metrics. Unlike outcome-based metrics (profit, accuracy) that evaluate results, or process-compliance metrics (checklists, audits) that evaluate procedural adherence, process-native metrics evaluate the quality of reasoning within the decision process itself using counterfactual criteria.


The Epistemic Signal Ratio (ESR) is the first instantiation of this class. It measures:


\text{ESR} = \frac{\text{Decision-Relevant Signal}}{\text{Total Documented Reasoning}} \quad \text{with} \quad 0 \leq \text{ESR} \leq 1


ESR operationalizes relevance through counterfactual perturbation: an information unit is relevant if its removal changes the ranking of decision alternatives. This definition shifts evaluation from subjective importance to decision impact—a distinction absent in prior literature.


1.4 Contributions and Roadmap


This paper makes five major contributions:


1. Measurement Theory: Introduces a formalized class of counterfactually grounded metrics, with ESR as its first instantiation, establishing it as a partial sufficient statistic for manifest reasoning efficiency conditional on a closed information set.

2. Decision Theory: Provides a reduced-form micro-foundation grounded in diminishing marginal cognitive returns, linking ESR to bounded rationality and information overload.

3. Causal Inference: Establishes identification through exogenous documentation shocks (LATE) with explicit first stage, demonstrating that ESR can be used as a causal estimand under quasi-experimental conditions.

4. Epistemic Integrity: Extends beyond efficiency to measure what organizations fail to consider (Latent Signal, Epistemic Blindspot), with latent signal defined relative to ex post validation rather than pure speculation, proving Proposition 4 (Groupthink Amplification).

5. Strategic Analysis: Shows that adding noise is a dominated strategy (ESR is incentive‑compatible with respect to noise addition), though selection manipulation remains possible (ESR is not fully strategy‑proof). This places ESR in the class of partially robust governance metrics.


The paper proceeds as follows. Section 2 establishes the counterfactual foundation. Section 3 distinguishes ESR from existing metrics. Section 4 provides reduced-form micro-foundations. Sections 5-9 present formal theorems with appropriate qualifications. Sections 10-17 elaborate comparative statics, testable predictions, and domain generalizability. Sections 18-31 detail measurement protocols. Sections 32-34 illustrate empirical feasibility. Sections 35-43 introduce Latent Signal and Epistemic Blindspot. Section 44 discusses limitations, and Section 45 concludes.


---


2. THEORETICAL FOUNDATIONS


2.1 Core Claim (Qualified)


“We show that existing governance systems cannot rigorously identify decision-relevant reasoning under counterfactual criteria.”


This claim is more precise than stating they are entirely “non-identifiable.” Existing systems may capture some aspects of reasoning, but they lack the counterfactual test needed to establish whether documented reasoning actually influenced the final decision.


2.2 Counterfactual Decision Impact


Unlike existing content-analysis approaches that rely on subjective importance ratings or frequency counts, ESR defines relevance strictly in counterfactual decision-theoretic terms: whether removing an information unit changes the ranking of alternatives.


Definition 1 (Decision-Relevant Signal)

An information unit i is a decision-relevant signal if:


\text{Relevance}(i) = 1 \ \text{if} \ \text{Ranking}(\text{Alternatives} \mid \text{Information}) \neq \text{Ranking}(\text{Alternatives} \mid \text{Information} \setminus \{i\})


This shifts the unit of evaluation from importance to decision impact—a distinction absent in prior literature.


Qualification on Binary Relevance: The definition uses a binary change in ranking. In practice, relevance may be continuous; for instance, removing information might not change the ordering but could reduce confidence in the chosen alternative or shift the threshold for acceptance. We treat binary relevance as a conservative lower bound of epistemic signal—information that flips ranking is certainly relevant, while information that only shifts confidence may still matter. Thus, ESR identifies a strict subset of epistemic signal (decision-pivotal signal) rather than the full set of decision-relevant reasoning. Future extensions may consider continuous relevance based on ranking distance or confidence shifts, but the binary definition already provides a tractable and defensible starting point that yields rich theoretical and empirical implications.


2.3 Classifying Governance Metrics


Metric Class Focus Examples

Outcome-based Results Profit, accuracy, impact

Process-compliance Procedural adherence Checklists, audit compliance

Process-native epistemic (formalized & counterfactually grounded) Reasoning quality ESR


2.4 Positioning Statement


“We introduce a formalized and counterfactually grounded class of metrics: process-native epistemic metrics. ESR is the first instantiation of this class.”


---


3. DISTINGUISHING ESR FROM EXISTING METRICS


Existing Metric Why ESR Differs

Readability ESR measures decision relevance, not ease of reading

Complexity ESR measures marginal impact, not abstract complexity

Information volume ESR is a ratio, not an absolute count; long documents can have high or low ESR

Expert judgment ESR uses structured counterfactual evaluation, not subjective ratings

Shapley value / feature importance Shapley measures marginal contribution to payoff (cardinal); ESR measures change in decision ranking (ordinal)—fundamentally non-equivalent


“ESR is not a value allocation method but a decision-perturbation metric, making it conceptually orthogonal to cooperative game-theoretic attribution.”


---


4. MICRO-FOUNDATIONS: REDUCED-FORM REPRESENTATION OF DIMINISHING MARGINAL COGNITIVE RETURNS


4.1 Motivation


Our micro-foundation is not derived from a fully specified cognitive process but adopts a reduced-form representation consistent with diminishing marginal cognitive returns. This approach captures the key trade-off between information acquisition and processing capacity without requiring a complete model of cognitive architecture, and it is widely used in the bounded rationality literature (Simon, 1955; Conlisk, 1996).


4.2 Reduced-Form Model


Consider a rational agent facing a trade-off between information acquisition and cognitive costs:


S(I) = \alpha \ln(1 + I)


c(I) = \frac{\gamma}{2} I^2


Where:


· S(I) = signal extracted from total information I

· \alpha = extraction efficiency (captures cognitive capacity)

· \gamma = marginal processing cost

· \theta = value per signal


Utility function:


U(I) = \theta S(I) - c(I) = \theta \alpha \ln(1 + I) - \frac{\gamma}{2} I^2


“We adopt a reduced-form representation consistent with diminishing marginal cognitive returns, rather than deriving from a fully specified cognitive process.”


4.3 ESR as Function of Information Volume


\text{ESR}(I) = \frac{S(I)}{I} = \frac{\alpha \ln(1 + I)}{I}


4.4 Boundary Insights


\lim_{I \to 0} \text{ESR}(I) = \alpha


\lim_{I \to \infty} \text{ESR}(I) = 0


Interpretation:


· At very low information volume, all information appears relevant (ESR → α)

· At very high information volume, dilution dominates (ESR → 0)


4.5 Optimal Information Volume


First-order condition:


\frac{\theta \alpha}{1 + I^*} = \gamma I^*


---


5. THEOREM 1: INFORMATION OVERLOAD AND ESR DECLINE


Theorem 1 (Information Overload and ESR Decline)

Under the reduced-form micro-foundation with S(I) = \alpha \ln(1 + I), ESR is strictly decreasing in I for I > I_{\max}, where I_{\max} solves \frac{d}{dI}[\ln(1+I)/I] = 0.


Proof:


\frac{d}{dI}\left(\frac{\ln(1+I)}{I}\right) = \frac{\frac{I}{1+I} - \ln(1+I)}{I^2} < 0 \quad \text{for} \quad I > I_0


Since \ln(1+I) > \frac{I}{1+I} for I > 0, the derivative is negative beyond a threshold.


Implication: Beyond a certain point, adding more reasoning reduces the efficiency of decision-relevant signal extraction, mechanically lowering ESR.


---


6. THEOREM 2: IDENTIFIABILITY UNDER PANEL AGGREGATION


Theorem 2 (Identifiability under Panel Aggregation)

Under independent panel evaluation with bounded classification error \epsilon < 0.5, the ESR estimator \widehat{\text{ESR}} converges in probability to the true latent ESR as K \to \infty.


Measurement Error Model:


\text{Relevance Score}(i) = R_i^* + u_i


where u_i i.i.d., mean zero, \text{Var}(u_i) = \epsilon(1-\epsilon).


Proof:

\widehat{\text{ESR}} = \frac{1}{N} \sum_{i=1}^{N} \text{Relevance Score}(i) = \text{ESR}_{\text{true}} + \bar{u}. By Law of Large Numbers, \bar{u} \xrightarrow{p} 0 as K \to \infty.


Qualification: The theorem assumes independence or weak dependence of panel judgments. In practice, panelists may share systematic biases (e.g., groupthink), which can violate this assumption. Our later analysis (Proposition 4) addresses such scenarios explicitly.


---


7. THEOREM 3: CAUSAL IDENTIFICATION


Theorem 3 (Causal Identification under Exogenous Documentation Shocks)


First Stage:


\text{ESR} = \pi_0 + \pi_1 Z + \eta \quad \text{with} \quad \pi_1 \neq 0


Second Stage (Wald Estimator):


\beta_1 = \frac{\text{Cov}(Q, Z)}{\text{Cov}(\text{ESR}, Z)} = \text{LATE} = \mathbb{E}[Q(1) - Q(0) \mid \text{Compliers}]


Where:


· Compliers: decision units whose ESR changes due to documentation shock Z

· Always-takers: units with high documentation regardless of Z

· Never-takers: units with low documentation regardless of Z


Examples of plausible instruments Z in governance settings:


· Staggered adoption of standardized decision documentation protocols across organizational units

· Externally mandated reporting requirements (e.g., donor-imposed standards, regulatory transparency rules)

· Introduction of AI-assisted documentation tools that alter the structure of recorded reasoning


Interpretation: The first stage ensures Z genuinely moves ESR (relevance condition). The estimand is the LATE for compliers—those whose reasoning structure is affected by the documentation shock.


---


8. DYNAMIC IMPLICATION: DECISION STABILITY


Prediction:


\text{Pr}(\text{Ranking change}_{t+1}) = \delta_0 - \delta_1 \text{ESR}_t \quad \text{with} \quad \delta_1 > 0


Or, in variance form:


\text{Var}(\text{Ranking}_{t+1}) = \sigma^2(\text{ESR}_t), \quad \frac{d\sigma^2}{d\text{ESR}} < 0


Interpretation: Higher ESR reduces the probability of future ranking changes—decisions are more robust to new information.


---


9. MEASUREMENT ERROR AND ATTENUATION BIAS


Attenuation Bias:


\hat{\beta}_1^{\text{OLS}} \rightarrow \lambda \beta_1 \quad \text{with} \quad 0 < \lambda < 1


Interpretation: Classical measurement error in ESR biases the OLS coefficient toward zero. If estimates remain significant despite this bias, the true effect is likely larger than estimated.


---


10. STRATEGIC MANIPULATION: PARTIALLY ROBUST GOVERNANCE METRICS


10.1 Adding Noise is a Dominated Strategy


Consider an agent who can add noise N (information units that do not change ranking):


\text{ESR}(N) = \frac{S}{S + N}


Result:


\frac{\partial \text{ESR}}{\partial N} = -\frac{S}{(S+N)^2} < 0


Adding noise strictly decreases ESR. Thus, ESR is incentive‑compatible with respect to noise addition: no agent who values ESR positively will add noise.


10.2 Selection Manipulation Remains Possible


However, agents can manipulate ESR through other channels:


· Selective documentation: omitting relevant dissent or counter-evidence (reducing denominator while keeping signal unchanged)

· Framing manipulation: presenting the same information in ways that appear more decisive

· Gaming the panel: influencing panelists’ judgment through persuasion or selection


These forms of manipulation are not ruled out by the above analysis. Therefore, ESR is not fully strategy‑proof.


“This places ESR in the class of partially robust governance metrics—robust to additive noise but vulnerable to selection and framing manipulation.”


---


11. FORMAL COMPARATIVE STATICS


From the first-order condition \frac{\theta \alpha}{1 + I^*} = \gamma I^*:


Parameter Derivative Sign Interpretation

Extraction efficiency \alpha \frac{\partial I^*}{\partial \alpha} > 0 + Higher capacity → more information processed

Value per signal \theta \frac{\partial I^*}{\partial \theta} > 0 + Higher incentives → more information

Cognitive cost \gamma \frac{\partial I^*}{\partial \gamma} < 0 - Higher cost → less information


Effects on optimal ESR:


\frac{\partial \text{ESR}^*}{\partial \alpha} > 0,\quad \frac{\partial \text{ESR}^*}{\partial \theta} > 0,\quad \frac{\partial \text{ESR}^*}{\partial \gamma} < 0


---


12. ESR AS A PARTIAL SUFFICIENT STATISTIC


“ESR can be interpreted as a partial sufficient statistic for manifest reasoning efficiency conditional on a closed information set.”


Under the assumption that all decision-relevant information is captured in the documented reasoning (i.e., no latent signal), ESR captures all relevant variation in the efficiency of processing that information. However, as we show in Sections 35-43, latent signal exists and matters. Therefore, ESR is not sufficient for overall decision quality—it is a partial sufficient statistic for the efficiency of manifest reasoning, but it must be complemented with measures of signal completeness (e.g., Epistemic Blindspot) to fully assess reasoning quality.


“ESR is sufficient for manifest reasoning efficiency conditional on a closed information set, but it is not sufficient for decision quality due to latent signal (Section 20).”


---


13. ESR VS SHAPLEY VALUE: FUNDAMENTAL NON-EQUIVALENCE


Aspect Shapley Value ESR

Domain Payoff (cardinal) Preference ranking (ordinal)

Invariance Monotonic payoff transformations Monotonic ranking transformations

Information needed Payoff values Ranking comparisons


“Shapley value is invariant to monotonic transformations of payoff, while ESR is invariant to monotonic transformations of preference rankings—making them fundamentally non-equivalent objects.”


“ESR is not a value allocation method but a decision-perturbation metric, making it conceptually orthogonal to cooperative game-theoretic attribution.”


---


14. COMPETING PREDICTIONS: SHARP TESTABLE RESTRICTIONS


Empirical Model:


Q = \beta_0 + \beta_1 \text{ESR} + \beta_2 \text{ESR}^2 + \varepsilon


Framework Empirical Restriction

ESR Theory \beta_1 > 0,\ \beta_2 < 0

Readability \beta_2 = 0 (monotonic positive)

Information volume \beta_2 = 0 (monotonic positive)

Complexity \beta_1 < 0,\ \beta_2 = 0 (monotonic negative)


---


15. LIMIT CASE ANALYSIS


Limit Case ESR Interpretation

All units relevant S = N, \text{ESR} = 1 Maximum efficiency

No signal S = 0, \text{ESR} = 0 No decision-relevant content

Noise → ∞ \lim_{N \to \infty} \frac{S}{S+N} = 0 ESR collapse

Information → 0 \lim_{I \to 0} \text{ESR}(I) = \alpha All information appears relevant

Information → ∞ \lim_{I \to \infty} \text{ESR}(I) = 0 Dilution dominates


---


16. EMPIRICAL TEST MAPPING


Prediction Empirical Design Identification

Noise ↑ → ESR ↓ External documentation shocks First stage: \pi_1 \neq 0; IV: \beta_1 = \text{LATE}

ESR–Q inverted-U Panel dataset \beta_1 > 0,\ \beta_2 < 0

Capacity ↑ → ESR* ↑ Training, AI assistance Before-after, DiD

Manipulation via selection Panel evaluation with fixed selection criteria Need additional controls

ESR ↑ → stability ↑ Panel longitudinal \delta_1 > 0


---


17. DOMAIN GENERALIZABILITY


Domain Example Applications

Public policy Regulatory impact analysis, strategic policy decisions

Corporate governance Board decisions, strategic investment

AI alignment Reasoning trace of generative AI models

Non-profit management Fund allocation, social programs


---


18. MEASUREMENT PROTOCOL


18.1 Unit of Analysis: Minimal Reasoning Unit


i \in \mathcal{U} = \text{set of minimal independently evaluable reasoning units}


Operational Rules:


1. Separability: Can be removed without destroying overall logical structure

2. Evaluative independence: Has a claim that can be independently assessed

3. Counterfactual testability: Can be tested in ranking framework


18.2 Reliability: Unitization and Relevance


\text{IRR}_{\text{unitization}} = \text{Krippendorff’s Alpha (unit segmentation)}


\text{IRR}_{\text{relevance}} = \text{Krippendorff’s Alpha (relevance scores)}


Interpretation: \geq 0.80 = reliable; 0.67-0.80 = acceptable; <0.67 = revision needed.


18.3 Relevance Scoring


\text{Relevance Score}(i) = \frac{1}{K} \sum_{k=1}^{K} \mathbb{I}_{\text{ranking changes according to panelist } k}


18.4 ESR Estimation with Uncertainty


\text{ESR}_{\text{true}} = \frac{1}{N} \sum_{i=1}^{N} \text{Relevance Score}(i)


\text{SE}_{\text{ESR}} = \frac{\text{StdDev}(\text{Relevance Score}(i))}{\sqrt{N}}


\text{CI}_{95\%} = \text{ESR}_{\text{true}} \pm 1.96 \times \text{SE}_{\text{ESR}}


18.5 Weighted ESR


\text{ESR}_{\text{weighted}} = \frac{\sum_{i=1}^{N} W_i \cdot R_i}{\sum_{i=1}^{N} W_i}


Category Weight (W_i)

Critical assumptions 3

Evidence 3

Substantive alternatives 2

Documented dissent 2

Minor supporting arguments 1


18.6 Tiered Implementation


Tier Method Cost Use Case

Tier 1: ESR Lite \text{ESR}_{\text{lite}} = \frac{\text{arguments used}}{\text{total arguments}} Low NGOs, quick pilots

Tier 2: Internal BCEV+ Internal cross-unit panel Medium Ministries, mid-sized corporations

Tier 3: Full BCEV+ External panel with uncertainty estimation High International organizations, accreditation


---


19. INTERPRETATION AND FAILURE THRESHOLD


ESR Level Condition Risk

< 0.20 Noise dominant Signal dilution

0.20 – 0.39 Much irrelevant information Cognitive overload

0.40 – 0.80 Signal dominant Maximum quality

0.90 Over-filtering Loss of nuance


---


20. EXTENDING BEYOND EFFICIENCY: LATENT SIGNAL AND EPISTEMIC BLINDSPOT


20.1 The Missing Dimension


ESR measures how efficiently organizations process manifest information. But what about information that should be considered but never enters deliberation? This is the domain of Latent Signal (S_L) and Epistemic Blindspot (EB).


Definition 2 (Latent Signal)

Information that objectively has counterfactual power to change decision ranking but is absent from documented reasoning or systematically ignored by decision-makers. Latent signal is defined relative to ex post validation or external domain knowledge, not purely speculative counterfactuals. Crucially, latent signal is not defined by realized outcomes alone, but by information that was ex ante available and ex post validated, excluding hindsight‑only constructs. For instance, a warning that a project would fail that was ignored and later proved correct qualifies as latent signal; a hypothetical argument that might have changed the decision but was never proposed does not.


Sources of latent signal:


· Political pressure or organizational culture (groupthink)

· Epistemic capture by dominant interests

· Cognitive limitations or selection bias

· Unknown unknowns (detectable only through systematic adversarial probing)


Definition 3 (Epistemic Blindspot)


EB = \frac{S_L}{S_M + S_L}


where S_M = manifest signal (signal present in documentation).


Definition 4 (Signal Completeness)


SC = 1 - EB


20.2 Adjusted Decision Quality


Q_{adj} = f(ESR, SC)


True decision quality depends not only on how efficiently we process available information (ESR), but also on how complete that information is (SC).


20.3 The Four Quadrants of Epistemic Health


ESR EB Condition Risk

High High Efficient but blind Missed forest for the trees; sudden crisis

Low Low Comprehensive but messy Analysis paralysis; slow decisions

Optimal Low Ideal Efficient and complete; long-term stability

Low High Chaotic and blind Total failure


---


21. PROPOSITION 4: GROUPTHINK AMPLIFICATION


Proposition 4 (Groupthink Amplification)

As groupthink (\phi) increases, the observed ESR may stay constant or even increase (due to artificial consensus), but the Latent Signal (S_L) increases exponentially, leading to a catastrophic drop in Q_{adj}.


Mechanism with a stylized functional form:

Let S_L(\phi) = S_0 e^{k\phi} where S_0 is baseline latent signal and k > 0 captures the amplification rate of ignored information under groupthink. Observed ESR, defined using only manifest signal S_M, remains stable or may even rise because consensus eliminates perceived dissent (reducing denominator in ESR calculation). Meanwhile, latent signal grows exponentially. When the ignored information eventually surfaces, the collapse in decision quality is sudden and severe.


Implications: High ESR does not guarantee safety. If EB is high, organizations are time bombs waiting to explode. This explains:


· Major project failures undetected beforehand

· Corporate scandals that “suddenly” emerge

· Policy crises that minority voices had warned about


---


22. DETECTING LATENT SIGNAL (ADVERSARIAL AUDIT)


Since S_L is “invisible” (absent from documentation), detection requires structured adversarial methods:


1. Red Teaming: Independent external panel searches for information not mentioned in documentation but relevant to the domain, using domain expertise and external benchmarks.

2. Counter-Ranking Test: Present new information to decision-makers; if ranking changes drastically, that information was S_L.

3. Dissent Gap: Calculate difference between suppressed minority opinions and final decision.


Epistemic Boundary: Latent signal is not purely speculative; it is validated ex post (e.g., the information later proved correct) or grounded in established domain knowledge. This prevents conflating hidden signals with hypothetical alternatives that may never have been relevant.


---


23. REVISED COMPARATIVE STATICS


Parameter Effect on ESR Effect on EB Effect on Quality (Q)

Psychological Safety Decreases (more discussion) Decreases sharply Increases

Groupthink (\phi) Increases (appears aligned) Increases sharply Decreases drastically

AI Auditor Stable Decreases Increases


---


24. LIMITATIONS AND FUTURE RESEARCH


Limitation Mitigation / Future Work

Unit analysis ambiguity Formal Minimal Reasoning Unit; separate IRR for unitization

Panel subjectivity and dependence Theorem 2 requires independence/weak dependence; future work on correlated errors

Endogeneity Theorem 3: causal identification with exogenous shocks; requires valid instruments

Measuring S_L Adversarial audit; acknowledged as estimation, not absolute measure; latent signal defined relative to ex post validation, excluding hindsight‑only constructs

Audit fatigue Need research on optimal frequency and integration with regular operations

Epistemic paranoia Need satisficing criteria for when to stop searching

Power asymmetry Need explicit integration with power structures (future work)

Selection manipulation Current analysis shows adding noise is dominated; manipulation via selection remains open; ESR is incentive‑compatible with respect to noise addition but not fully strategy‑proof


---


25. EMPIRICAL ILLUSTRATION: IDENTIFICATION FEASIBILITY


Dataset: N = 300, ESR ∼ Uniform(0.1, 0.9), Q = 0.5 + 1.5\,\text{ESR} - 1.8\,\text{ESR}^2 + \varepsilon


Results:


Variable Coefficient Std. Error p-value

ESR 1.48 0.21 < 0.001

ESR² -1.72 0.18 < 0.001


\text{ESR}^* \approx 0.43,\quad R^2 = 0.58


Interpretation: The empirical illustration establishes identification feasibility. The next empirical step is quasi-experimental validation using externally induced variation in documentation protocols.


---


26. DIAGRAM OF ESR MEASUREMENT


```

┌─────────────────────────────────────────────────────────────────────────────┐

│  Information Supply + Cognitive Filtering + Organizational Routines        │

│                              │                                             │

│                              ▼                                             │

│  ┌───────────────────────────────────────────────────────────────────────┐ │

│  │ STEP 1: UNITIZATION → Minimal Reasoning Units + IRR_unitization      │ │

│  └───────────────────────────────────────────────────────────────────────┘ │

│                              │                                             │

│                              ▼                                             │

│  ┌───────────────────────────────────────────────────────────────────────┐ │

│  │ STEP 2: PANEL EVALUATION (BCEV+) → Relevance Score(i) + IRR_relevance│ │

│  └───────────────────────────────────────────────────────────────────────┘ │

│                              │                                             │

│                              ▼                                             │

│  ┌───────────────────────────────────────────────────────────────────────┐ │

│  │ STEP 3: ESR ESTIMATION → ESR_true ± SE_ESR, CI_95%                   │ │

│  └───────────────────────────────────────────────────────────────────────┘ │

│                              │                                             │

│                              ▼                                             │

│  ┌───────────────────────────────────────────────────────────────────────┐ │

│  │ STEP 4: OUTCOME ANALYSIS → Q = α + β₁ ESR + β₂ ESR² + ε              │ │

│  │ Sharp restriction: β₂ < 0                                            │ │

│  └───────────────────────────────────────────────────────────────────────┘ │

└─────────────────────────────────────────────────────────────────────────────┘

```


---


27. SIMULATION ILLUSTRATION


```

Decision Quality (Q)

      ↑

  1.0 |        *

      |       * *

  0.8 |      *   *

      |     *     *

  0.6 |    *       *

      |   *         *

  0.4 |  *           *

      | *             *

  0.2 |*               *

      |____________________→ ESR

       0    0.2   0.4   0.6   0.8   1.0

                     ↑

                 ESR* ≈ 0,43

```


---


28. CONCLUSION


28.1 Summary of Contributions


1. Counterfactual perturbation – Redefines relevance as decision impact rather than subjective importance; fundamentally distinct from Shapley (ordinal vs cardinal invariance). ESR identifies a strict subset of epistemic signal (decision-pivotal signal) rather than the full set of decision-relevant reasoning.

2. Reduced-form micro-foundation – S(I) = \alpha \ln(1+I), justified by diminishing marginal cognitive returns rather than derived from a fully specified cognitive process

3. Theorem 1 (Information Overload) – Explicit derivation of ESR decline beyond optimal volume

4. Theorem 2 (Identifiability) – Consistent estimation under panel aggregation, with qualification about independence

5. Theorem 3 (Causal Identification) – LATE for compliers with explicit first stage and concrete instrument examples

6. Dynamic implication – \text{Pr}(\text{Ranking change}_{t+1}) = \delta_0 - \delta_1 \text{ESR}_t

7. Attenuation bias – \hat{\beta}_1^{\text{OLS}} \rightarrow \lambda \beta_1

8. Strategic analysis – Adding noise is dominated (incentive‑compatible with respect to noise addition); selection manipulation remains possible (not fully strategy‑proof). This places ESR in the class of partially robust governance metrics.

9. Partial sufficient statistic – ESR summarizes manifest reasoning efficiency conditional on a closed information set; it is not sufficient for decision quality due to latent signal

10. Sharp testable restrictions – \beta_1 > 0,\ \beta_2 < 0 distinguishing ESR from alternatives

11. Epistemic Blindspot – Extends framework to measure what organizations fail to consider, with latent signal defined relative to ex post validation, excluding hindsight‑only constructs

12. Proposition 4 (Groupthink Amplification) – High ESR with high EB is a time bomb; stylized form S_L(\phi) = S_0 e^{k\phi} illustrates exponential growth of ignored information


28.2 Core Claims (Qualified)


“We show that existing governance systems cannot rigorously identify decision-relevant reasoning under counterfactual criteria.”


“We introduce a formalized and counterfactually grounded class of metrics: process-native epistemic metrics. ESR is the first instantiation of this class.”


“ESR makes epistemic accountability measurable, complementing existing governance systems that focus on outcomes and compliance.”


“ESR is incentive‑compatible with respect to noise addition, but not fully strategy‑proof. This places ESR in the class of partially robust governance metrics.”


28.3 Final Statement


“ESR does for reasoning what productivity metrics did for labor: it transforms an invisible process into a measurable, optimizable, and governable variable—while recognizing that no single metric can capture all dimensions of reasoning quality.”


---


29. REFERENCES


· Angrist, J. D., & Pischke, J.-S. (2009). Mostly Harmless Econometrics. Princeton University Press.

· Conlisk, J. (1996). Why bounded rationality? Journal of Economic Literature, 34(2), 669–700.

· Grimmelikhuijsen, S., Jilke, S., Olsen, A. L., & Tummers, L. (2017). Behavioral public administration: Combining insights from public administration and psychology. Public Administration Review, 77(1), 45–56.

· Kahneman, D. (2011). Thinking, Fast and Slow. Farrar, Straus and Giroux.

· Kahneman, D., & Tversky, A. (1979). Prospect theory: An analysis of decision under risk. Econometrica, 47(2), 263–291.

· Krippendorff, K. (2018). Content Analysis: An Introduction to Its Methodology (4th ed.). Sage.

· Simon, H. A. (1947). Administrative Behavior. Macmillan.

· Simon, H. A. (1955). A behavioral model of rational choice. Quarterly Journal of Economics, 69(1), 99–118.

· Shapley, L. S. (1953). A value for n‑person games. Annals of Mathematics Studies, 28, 307–317.


---


30. APPENDICES (Indicative)


· Appendix A: Coding Manual for Minimal Reasoning Unit

· Appendix B: Scoring Rubric for BCEV+ Panel

· Appendix C: Decision Traceability Log Template

· Appendix D: IRR Results (Unitization and Relevance)

· Appendix E: Dataset Synthesis Code (Python/R) for Replication

· Appendix F: Full Regression Outputs with Robust Standard Errors

· Appendix G: Adversarial Audit Checklist for Latent Signal Detection


---


Accountability‑Based Universal Wisdom and Trust

Cross‑Sector Pre‑Decision Governance Translator


Final Manuscript · March 2026


License: CC BY-NC-SA 4.0

Contact: tpapgtk@gmail.com

Archive: https://abuwt.blogspot.com