The Epistemology of Public Audit: From Rituals of Verification to Epistemic Accountability
Abu Abdurrahman, M.H.
Abstract
Public audit systems have expanded dramatically in scope and sophistication over the past four decades. Yet a persistent anomaly remains: projects and policies that satisfy all procedural and financial compliance requirements continue to produce weak outcomes—cost overruns, delays, and failure to deliver public value. This paper terms this phenomenon output-compliant but outcome-weak governance. Drawing on Power's (1997) critique of the audit society, the paper traces the trajectory of post-Power governance reforms—performance audit expansion, risk-based governance, evidence-based policymaking, anticipatory governance, and behavioral public administration—and argues that while these developments successfully widened the evaluative reach of governance, they did not alter its fundamental architecture: the unit of accountability remained outputs, compliance, and risk, rather than the procedural robustness of upstream deliberation. The paper introduces the concept of institutionalized legibility bias—the tendency of organizations to optimize for what accountability systems visibly reward—as a theoretical mechanism explaining the systematic neglect of upstream reasoning quality. Drawing on Scott's (1998) analysis of state legibility, Goodhart's (1984) law on metric fixation, and the performativity literature (Strathern, 2000; Espeland & Sauder, 2007; Muller, 2018), it argues that accountability systems not only evaluate organizational behavior but actively shape it toward institutionally visible dimensions at the expense of epistemically significant but less visible ones. Organizations optimize for auditable defensibility—the capacity to survive audit scrutiny—rather than for the cognitive robustness of their decisions, because audit risk, legal exposure, and political blame are more immediate organizational imperatives than long-term outcome quality. This mechanism is proposed as the paper's central explanatory engine: it is auditable defensibility that drives organizations to invest in what accountability systems reward and to neglect what they do not. It proposes epistemic accountability as a governance orientation, operationalized through decision process accountability mechanisms, to address this gap. The central contribution is ontological rather than prescriptive: epistemic accountability is distinguished from procedural rationality, deliberative governance, and risk governance not by advocating better deliberation, but by shifting the object of accountability from institutional outputs to the reconstructability of institutional reasoning. A dedicated section demonstrates why epistemic accountability is not reducible to procedural accountability: the former verifies whether required steps occurred; the latter verifies whether reasoning structures became institutionally reconstructable—a different accountability ontology. The paper's primary claim is ontological and institutional, not deterministic-performative. The central argument is that the recoverability of reasoning constitutes a missing dimension of accountability, regardless of whether such recoverability consistently improves outcomes. Instrumental benefits remain contingent empirical hypotheses, not foundational premises of the framework. The paper acknowledges that reasoning visibility itself can become politically captured, and that structural inequalities in who can contest decisions must be addressed. It explicitly addresses the framework's relationship to power asymmetries, bureaucratic domination, and epistemic hierarchy. It concludes with an empirical validation agenda centered on a falsifiable proposition.
Keywords: epistemic accountability, public audit, rituals of verification, decision process accountability, institutionalized legibility bias, auditable defensibility, upstream deliberation, governance theory
---
1. Introduction
Public audit institutions have never been more capable. Supreme Audit Institutions (SAIs) and internal audit bodies across jurisdictions now deploy increasingly sophisticated methodologies—performance auditing, risk-based auditing, data analytics, and real-time monitoring—to ensure that public funds are spent in compliance with rules and that programs achieve their intended outputs. The expansion of audit mandates over the past four decades, described by Power (1997) as the rise of the audit society, has transformed audit from a narrow financial control function into a comprehensive governance mechanism.
Yet a persistent anomaly challenges this narrative of progressive improvement. Projects that are fully compliant with procurement rules, that absorb their budgets, that produce their physical outputs on schedule, and that receive unqualified audit opinions nonetheless fail to generate meaningful public value. Infrastructure projects experience cost overruns of 50–200% despite passing multiple layers of review (Flyvbjerg, 2017). Public facilities—markets, sports complexes, cultural centers—stand underutilized years after their completion. Social programs deliver their administrative outputs while the outcomes they were designed to improve remain stagnant. This paper terms this phenomenon output-compliant but outcome-weak governance.
The recurrence of this pattern across sectors, jurisdictions, and levels of government suggests that the problem is not one of inadequate compliance or insufficient auditing intensity. Rather, it points to a structural blind spot in the epistemological foundations of contemporary governance: the systematic neglect of the quality of deliberation that precedes formal decision commitment.
The dominant accountability paradigm operates on a retrospective logic. Financial audits examine whether funds were spent in accordance with appropriations after the fact. Compliance audits verify whether procedures were followed after decisions were executed. Performance audits assess whether outputs were achieved after programs were implemented. Each modality is essential, but each arrives on the scene after the critical upstream decisions—about what problem to address, which assumptions to adopt, which alternatives to consider, and which risks to accept—have already been locked into place.
This paper argues that this temporal and conceptual gap constitutes an epistemic deficit in the architecture of public accountability. Governance systems excel at answering the question: "Were the rules followed?" They are considerably less equipped to answer: "Is the reasoning behind this decision recoverable, contestable, and reviewable independently of whether the outcome turned out well?" The distinction has direct consequences for whether public resources generate genuine value or merely produce procedurally impeccable failures.
The core contribution of this paper is to argue that what is missing is not a specific reform but an entire dimension of accountability. Contemporary governance architectures are institutionally legible—they render outputs, compliance, and financial regularity visible and reviewable. They are far less legible with respect to reasoning: the assumptions, alternatives, dissent, and stress-testing that constitute the deliberative foundation of decisions remain largely invisible to the formal accountability apparatus. This asymmetry is not accidental; it is produced by an institutionalized legibility bias—the systematic tendency of organizations to optimize for what accountability systems visibly reward, and to underinvest in what they do not.
The mechanism driving this bias is auditable defensibility—the organizational imperative to survive audit scrutiny. Organizations optimize for auditable defensibility rather than for the cognitive robustness of their decisions, because audit risk, legal exposure, and political blame are more immediate organizational imperatives than long-term outcome quality. A decision that is well-reasoned but poorly documented exposes the decision-maker to greater institutional vulnerability than a decision that is poorly reasoned but impeccably documented. The incentive structure rewards the latter and penalizes the former. It is auditable defensibility that explains why proceduralism persists, why organizations optimize documentation over deliberation, and why the invisibility of reasoning is not an oversight but a rational adaptation to existing accountability architectures.
The paper proposes epistemic accountability as a governance orientation designed to address this asymmetry. Its central ontological claim is that epistemic accountability is distinguished from related concepts—procedural rationality, deliberative governance, risk governance—not by advocating better deliberation, but by shifting the object of accountability from institutional outputs to the reconstructability of institutional reasoning. Deliberative governance evaluates the inclusiveness and legitimacy of discourse in democratic forums; epistemic accountability evaluates the recoverability and reviewability of reasoning under institutional decision conditions. The focus is not on improving cognitive quality but on establishing institutional conditions under which reasoning becomes recoverable, contestable, and traceable.
A critical distinction must be drawn: epistemic accountability is not simply procedural accountability with new vocabulary. This distinction is important enough to warrant dedicated treatment—which it receives in Section 4.2—but the essence is this: procedural accountability verifies whether required steps occurred; epistemic accountability verifies whether reasoning structures became institutionally reconstructable. The former asks, "Was the risk assessment completed?" The latter asks, "Were the assumptions that shaped the risk assessment documented, exposed to challenge, and made recoverable for subsequent review?" They are different accountability ontologies.
The paper's primary claim is ontological and institutional, not deterministic-performative. The central argument is that the recoverability of reasoning constitutes a missing dimension of accountability, regardless of whether such recoverability consistently improves outcomes. While there are plausible generative mechanisms through which deliberative visibility may enhance decision quality—specified in Section 4.5—these remain contingent empirical hypotheses, not foundational premises of the framework. The framework does not depend on the claim that visibility improves outcomes; it depends on the claim that invisibility represents an accountability deficit.
It is essential to clarify what epistemic accountability does not claim to do. Epistemic accountability does not audit cognition directly. It does not claim access to the internal mental states of decision-makers, nor does it purport to measure sincerity, intelligence, or wisdom. What it audits is far more modest: the recoverability, traceability, contestability, and exposure of assumptions in the documentary record of deliberation. It is concerned not with whether decision-makers thought well but with whether the institutional conditions existed for their reasoning to be reconstructed, challenged, and reviewed by others. This distinction is fundamental to the framework's epistemological defensibility and distinguishes it from impossible positivist ambitions to audit internal cognition.
The paper is positioned as a contribution to governance and accountability theory, using public audit as its empirical entry point. However, the orientation it proposes is not confined to audit institutions. Epistemic accountability can, in principle, be embedded within cabinet systems, judicial review, corporate boards, central banks, development banks, military planning, pandemic response, and climate adaptation governance—any institutional setting where consequential decisions are made under uncertainty.
The paper proceeds as follows. Section 2 critically examines the epistemological foundations of public audit, traces the trajectory of post-Power governance reforms, and introduces the concept of institutionalized legibility bias and its driving mechanism, auditable defensibility. Section 3 maps the specific epistemic gap. Section 4 proposes epistemic accountability, distinguishes it from adjacent concepts, provides the central argument for its irreducibility to procedural accountability, and positions the framework within governance theory. Section 5 outlines a preliminary operational architecture. Section 6 discusses non-claims, the potential pathologies of deliberation, governance trade-offs, the framework's relationship to power asymmetries, institutional resistance, recursive ritualization, the distinction between epistemic failure and political rationality, boundary conditions, and the empirical validation agenda. Section 7 concludes with a philosophical synthesis.
---
2. The Epistemology of Public Audit and Its Aftermath
2.1. The Audit Society and Its Discontents
Michael Power's (1997) The Audit Society provided the most sustained critique of the explosion of audit activity that began in the 1980s. Power's central argument was that audit had expanded far beyond its traditional financial domain into areas where the underlying constructs were inherently ambiguous—quality, effectiveness, value for money, governance. In these domains, audit did not simply measure pre-existing realities; it actively shaped what counted as legitimate organizational behavior, often in ways decoupled from substantive performance.
Power's key insight was that audit operates as a ritual of verification—a set of formalized procedures that produce comfort and legitimacy for external audiences rather than substantive insight into organizational quality. Organizations learn to be "auditable" by producing the documentation that audit demands, with negligible effects on the underlying activities being audited. Audit scores and compliance indicators improve while actual outcomes remain stagnant.
This critique has been developed in subsequent literature. Shore and Wright (2015) documented how audit cultures in universities reshape academic behavior toward "auditable" outputs at the expense of intellectual depth. Strathern (2000) explored the tyranny of transparency, showing how the demand for ever more visible indicators generates gaming and superficial compliance. Bovens (2007) distinguished between accountability as a virtue and as a mechanism, warning that an overemphasis on mechanistic accountability could undermine the substantive responsibility it was meant to serve.
Collectively, this literature demonstrates that the expansion of audit has not been accompanied by a proportionate improvement in the quality of public decisions. The rituals of verification are performed with increasing sophistication, while the reasoning that shapes the decisions being verified remains outside the accountability frame.
2.2. The Temporal Asymmetry of Governance
A central feature of the dominant accountability paradigm is its retrospective temporal orientation. Financial, compliance, and performance audits examine what has already occurred—funds spent, procedures followed, outputs delivered. This orientation is not a flaw; it is a defining feature. Accountability's legitimacy rests in part on its capacity to provide an objective account of past actions based on available evidence.
However, this creates a structural gap. The decisions whose consequences accountability examines were made earlier, under conditions of uncertainty, based on assumptions that may or may not have been explicitly articulated, alternatives that may or may not have been systematically explored, and dissenting views that may or may not have been documented. By the time accountability intervenes, these upstream processes are largely invisible.
This temporal asymmetry has two consequences. First, accountability is systematically least present at the stage where its potential contribution to decision quality could be greatest: before commitment, when assumptions are still testable and the costs of correction are low. Second, when accountability does identify failure, its findings arrive too late to prevent the specific failure under examination.
2.3. Institutionalized Legibility Bias: Accountability as Incentive Architecture
The retrospective, compliance-oriented nature of accountability interacts with organizational incentives in a way that systematically disadvantages upstream deliberation. This section develops the concept of institutionalized legibility bias as a theoretical mechanism explaining this dynamic.
The concept draws on Scott's (1998) analysis of how states make societies legible through standardized metrics, census categories, and cadastral maps—simplifications that render complex realities administratively manageable but systematically exclude what cannot be captured in those simplifications. The same dynamic operates within accountability systems: they render certain organizational behaviors institutionally legible while leaving others invisible.
This logic is reinforced by Goodhart's law: when a measure becomes a target, it ceases to be a good measure (Goodhart, 1984). Organizations adapt to accountability metrics by optimizing precisely what is measured, often at the expense of what is not. Strathern (2000) extended this insight to governance contexts, showing how transparency demands generate "tyranny" through indicator fixation. Espeland and Sauder (2007) demonstrated how rankings and metrics reshape organizational behavior toward reactivity—organizations responding to measures rather than to the underlying phenomena the measures purport to represent. Muller (2018) documented how metric fixation across sectors produces systematic gaming, short-termism, and neglect of unmeasurable dimensions of performance.
The result is what this paper terms institutionalized legibility bias: the systematic tendency of organizations to invest in what accountability systems visibly reward—financial regularity, procedural compliance, output documentation—while underinvesting in epistemically significant activities that lack institutional visibility. Assumption testing, alternative exploration, dissent facilitation, and stress-testing are precisely the kinds of activities that fall into this accountability blind spot. They are cognitively demanding but institutionally invisible; they improve decision quality but generate no audit credit; they expose vulnerabilities that organizations may prefer to keep implicit.
The driving mechanism behind this bias is auditable defensibility. Organizations optimize for auditable defensibility—the capacity to survive audit scrutiny—rather than for the cognitive robustness of their decisions, because audit risk, legal exposure, and political blame are more immediate organizational imperatives than long-term outcome quality. A decision that is well-reasoned but poorly documented exposes the decision-maker to greater institutional vulnerability than a decision that is poorly reasoned but impeccably documented. The incentive structure rewards the latter and penalizes the former. It is auditable defensibility that explains why proceduralism persists despite decades of critique, why organizations invest so heavily in documentation that may be substantively hollow, and why the invisibility of reasoning is not an oversight but a rational organizational adaptation. The logic is self-reinforcing: organizations produce what auditors can verify; auditors develop methodologies to verify what organizations produce; and the cycle entrenches an accountability architecture that systematically excludes upstream reasoning from institutional visibility.
This mechanism produces epistemic decoupling: a growing divergence between procedural accountability and substantive decision quality. Organizations learn to produce impeccably auditable decision processes that are procedurally complete but substantively hollow. The ritual of verification is performed, comfort is produced, and the underlying quality of reasoning remains unexamined.
This is not a failure of individual actors. It is a structural feature of an accountability architecture designed to govern financial probity and procedural compliance, not the cognitive quality of decision-making.
2.4. The Post-Power Governance Trajectory: Widening Evaluative Reach Without Altering Accountability Architecture
The quarter-century following Power's critique witnessed a series of governance reforms representing a sustained attempt to address the limitations of traditional audit. Each widened the evaluative reach of governance. Yet none fundamentally altered the unit of accountability: the architecture of reasoning itself remained outside formal review.
Performance auditing expanded audit's object from financial compliance to economy, efficiency, and effectiveness. This was a genuine advance, but it did not alter audit's temporal orientation. Auditors continued to assess whether targets were achieved, not whether the reasoning behind the targets was sound. Performance audit could detect execution inefficiency; it had no methodology for detecting the inadequacy of the reasoning that designed the execution.
Risk-based governance embedded risk assessment into organizational decision-making. This was, in principle, a move upstream: risk governance is forward-looking. Yet in practice, it focused on auditing identified risks—those formally registered and catalogued—not on auditing how risks were identified, which risks were excluded, or what framing assumptions shaped the assessment. Risk governance assumed the framing; it did not make the framing accountable.
Evidence-based policymaking sought to ground public decisions in rigorous empirical research. This was a significant epistemic advance, but it focused on evidence quality rather than reasoning process quality. Robust evidence could be deployed in service of a poorly framed problem, a narrow set of alternatives, or a decision process that systematically excluded dissenting perspectives. Good evidence does not, by itself, guarantee that the deliberation surrounding its use is recoverable or contestable.
Behavioral public administration demonstrated that public administrators are subject to systematic cognitive biases—overconfidence, confirmation bias, framing effects, groupthink (Grimmelikhuijsen et al., 2017; Moynihan, 2018). This literature provided a powerful diagnostic vocabulary but remained primarily a diagnostic enterprise. It identified cognitive vulnerabilities; it did not propose an institutional architecture for rendering deliberative quality systematically reviewable.
Anticipatory governance built institutional capacity for systematic foresight—scenario planning, horizon scanning, structured consideration of future risks (Guston, 2014; OECD, 2022). This was perhaps the most significant move upstream. Yet anticipatory governance remained primarily advisory. Decisions could be taken without systematic stress-testing of assumptions; foresight could be consulted or ignored at the discretion of decision-makers. It generated futures; it did not make the reasoning about those futures auditable.
Collectively, these five trajectories widened the evaluative reach of governance—from financial compliance to performance, to risk, to evidence, to cognition, to foresight. Yet they shared a common limitation: they expanded the object of evaluation without altering the architecture of accountability. Each generated new knowledge and new standards. None made the reasoning process itself—the assumptions, alternatives, dissent, and stress-testing that constitute the deliberative foundation of decisions—a systematic object of institutional accountability.
---
3. The Epistemic Gap: What Governance Does Not See
If the preceding analysis is correct, a significant proportion of governance failures can be attributed to epistemic deficits in upstream decision processes—deficits that are systematic and observable but fall outside the conceptual vocabulary and operational instruments of existing governance frameworks.
3.1. Unexamined Assumptions
Every decision rests on assumptions that, if incorrect, would fundamentally alter its expected consequences. Kahneman and Tversky's (1979) prospect theory demonstrated systematic underestimation of adverse outcome probabilities. Janis (1982) identified groupthink as a mechanism through which critical assumptions go unchallenged—not because they are correct, but because the social costs of challenging them are too high. March and Simon (1958) identified how organizational structures filter information in ways that systematically exclude uncomfortable assumptions.
In some cases, the failure to examine assumptions is not an oversight but a result of normalization of deviance—the gradual process identified by Vaughan (1996) in her study of the Challenger disaster, through which organizations come to accept what was once considered unacceptable through incremental acceptance of small deviations from standards. Assumptions that would have been questioned at an earlier stage become routinized background conditions, normalized into invisibility.
The practical consequence is that many public decisions proceed without systematic examination of their foundational assumptions. When these assumptions fail, the resulting shortfalls are attributed to "external factors" rather than to the upstream failure to test the assumptions that made the decision vulnerable.
3.2. Narrow Problem Framing
How a problem is framed largely determines the range of solutions considered (Simon, 1947). Existing governance systems assess whether the chosen solution was implemented in compliance with rules; they are far less equipped to assess whether the problem was framed adequately. The phenomenon of infrastructure bias—the systematic translation of complex public problems into physical infrastructure responses—illustrates this dynamic. Existing review mechanisms can verify whether a building was constructed to specification; they cannot evaluate whether the decision to construct it was based on an adequate diagnosis of the problem.
3.3. Unexplored Alternatives
Decision quality depends crucially on the range of alternatives considered. Many public decisions are characterized by an alternative deficit: the absence of systematic exploration of multiple options, with non-infrastructure alternatives frequently overlooked. Heuer's (1999) Analysis of Competing Hypotheses demonstrates how systematic consideration of alternatives can counteract confirmation bias, but this methodology has not been institutionalized within accountability architectures. As Perrow (1984) documented in his analysis of normal accidents, complex systems can produce catastrophic failures not because of individual errors but because of the tight coupling of system components that leaves no room for alternative pathways when something goes wrong. The failure to explore alternatives is not merely a cognitive lapse; it can be a structural feature of organizational design.
3.4. Suppressed Dissent
Dissenting views are a valuable resource for decision quality. Janis's (1982) work on groupthink demonstrated how the suppression of dissent can lead to catastrophic failures even among competent, well-intentioned decision-makers. Tetlock's (2005) research on expert political judgment demonstrated that decision-makers who consider a wider range of counter-arguments produce more accurate judgments. Yet existing governance frameworks have limited capacity to detect suppressed dissent. The absence of documented dissent is treated as consensus rather than a potential indicator of a deliberation deficit. Governance currently has no methodology for distinguishing genuine consensus from artificial consensus produced by the suppression of disagreement. Brunsson's (2002) concept of organizational hypocrisy is instructive here: organizations maintain formal structures and procedures that project rationality and consensus while informal practices, often involving substantive disagreement and political negotiation, operate beneath the surface. The formal record reflects the appearance of deliberation, not its reality.
---
4. Epistemic Accountability: A Governance Orientation
The preceding analysis suggests that existing accountability systems suffer from a structural gap: they evaluate decisions after they are made but provide no systematic mechanism for evaluating whether the reasoning that produced them is recoverable, contestable, and reviewable. This section introduces epistemic accountability as a governance orientation.
4.1. The Core Orientation and Its Ontological Distinction
Epistemic accountability is distinguished from related governance concepts by a specific ontological claim: its focus is not on improving the cognitive quality of deliberation—though this may be a desirable consequence—but on making deliberation institutionally reviewable independent of outcome correctness.
This distinction can be sharpened through contrast with proximate concepts. Procedural rationality (Simon, 1976) is a cognitive standard internal to decision-makers; epistemic accountability is an institutional condition external to them. Deliberative governance (Habermas, 1996) evaluates the inclusiveness and legitimacy of discourse in democratic forums; epistemic accountability evaluates the recoverability and reviewability of reasoning under institutional decision conditions. Risk governance evaluates whether risks have been identified and mitigated; epistemic accountability evaluates whether the process through which risks were identified, framed, and assessed is itself recoverable and contestable. Good governance is a broad normative ideal; epistemic accountability is a specific institutional dimension—the traceability of reasoning—that governance architectures may or may not incorporate.
A critical clarification must be made at the outset: epistemic accountability does not audit cognition directly. It does not claim access to the internal mental states of decision-makers, nor does it purport to measure sincerity, intelligence, or wisdom. What it audits is far more modest: the recoverability, traceability, contestability, and exposure of assumptions in the documentary record of deliberation. It is concerned not with whether decision-makers thought well but with whether the institutional conditions existed for their reasoning to be reconstructed, challenged, and reviewed by others. The object of evaluation is not the quality of thought but the visibility of the reasoning structure. This distinction is fundamental to the framework's epistemological defensibility and distinguishes it from impossible positivist ambitions to audit internal cognition.
The core claim, therefore, is not that epistemic accountability produces better decisions—a proposition requiring empirical validation—but that contemporary accountability architectures render upstream reasoning insufficiently recoverable, and that this represents a correctable institutional deficit.
A decision process exhibits greater deliberative visibility to the extent that it displays three characteristics:
· Traceability: critical assumptions are explicitly documented rather than implicitly accepted, so that the reasoning behind a decision can be reconstructed by an independent reviewer.
· Contestability: structured opportunities exist for dissenting views to be articulated, documented, and addressed before a decision is finalized, so that the range of perspectives considered is institutionally visible.
· Stress-testing: systematic examination of how a decision would perform under scenarios that differ from baseline assumptions is conducted and documented, so that the vulnerability of reasoning to foreseeable contingencies is made explicit.
Critically, deliberative visibility is evaluated independently of ex-post outcomes. A decision may exhibit high procedural traceability yet still fail due to unforeseeable exogenous shocks. The framework's objective is to ensure that the reasoning behind decisions is articulate, contestable, and recoverable—not that decisions are "correct."
4.2. Why Epistemic Accountability Is Not Reducible to Procedural Accountability
A sufficiently skeptical reviewer could argue that epistemic accountability is simply procedural accountability with new vocabulary—that requiring documentation of assumptions, alternatives, and dissent is ultimately just "more procedure." This section demonstrates why this is not the case. The distinction is not one of degree—more steps, better documentation—but one of kind: a different accountability object, a different temporal logic, and a different institutional function.
Procedural accountability verifies whether required steps occurred. Did the committee convene? Was the risk assessment completed? Were the procurement rules followed? The object of evaluation is the execution of mandated actions. Epistemic accountability verifies whether reasoning structures became institutionally reconstructable. Were the assumptions that shaped the risk assessment documented? Were the alternatives that were considered—and those that were rejected—made explicit? Was there a structured opportunity for dissenting views to be recorded? The object of evaluation is the recoverability of the reasoning process.
This distinction matters for five reasons.
First, the objects of accountability differ. Procedural accountability asks, "Was step X completed?" Epistemic accountability asks, "Can an independent reviewer reconstruct why this decision was made, on what assumptions, considering what alternatives, and in light of what disagreements?" A decision can satisfy all procedural requirements while remaining epistemically opaque. Conversely, a decision may deviate from standard procedure but be epistemically robust because the reasoning behind the deviation is fully documented and contestable.
Second, the evidentiary requirements differ. Procedural accountability relies on verification of actions—minutes of meetings, signed approvals, checklist completions. Epistemic accountability relies on the documentary record of deliberation—the articulation of assumptions, the exploration of alternatives, the documentation of dissent. The former confirms that the organization acted; the latter makes reconstructable how it thought.
Third, the temporal logic differs. Procedural accountability is overwhelmingly retrospective: it examines whether steps were followed after the fact. Epistemic accountability is structurally prospective: it operates before commitment, at the stage where assumptions can still be tested and alternatives remain available. This is not merely a temporal shift but a functional one—from evaluating the past to shaping the future.
Fourth, the institutional function differs. Procedural accountability protects organizations by demonstrating rule compliance. Epistemic accountability protects decisions by demonstrating reasoning robustness. The former provides a shield against accusations of procedural irregularity; the latter provides a shield against accusations of arbitrary or unexamined judgment. In the language of the Business Judgment Rule, procedural accountability addresses questions of compliance; epistemic accountability addresses questions of due care.
Fifth, and most fundamentally, the failure modes differ. Procedural accountability fails when rules are broken—when steps are skipped, approvals are forged, or documentation is incomplete. These failures are detectable within the existing accountability framework. Epistemic accountability fails when rules are followed but reasoning is hollow—when assumptions go unexamined, alternatives go unexplored, and dissent goes unexpressed, all while procedures are impeccably observed. These failures are systematically invisible to procedural accountability, and that is precisely the gap this framework addresses.
The practical implication is that procedural accountability can be fully satisfied while epistemic accountability is entirely absent. Indeed, the compliance trap identified in Section 2.3 describes precisely this condition: organizations that are procedurally impeccable yet epistemically vacuous. Adding more procedural requirements—more checklists, more approvals, more documentation mandates—cannot close this gap, because the gap is not in the quantity of procedure but in the object of accountability. Procedural accountability cannot make reasoning reconstructable for the same reason that a thermometer cannot measure humidity: it is the wrong instrument for the dimension being assessed.
This irreducibility argument is central to the paper's contribution. If epistemic accountability were merely procedural accountability with additional requirements, it would represent an incremental reform at best and redundant bureaucratic burden at worst. Because it operates on a different accountability dimension entirely—the visibility and recoverability of reasoning rather than the verification of actions—it represents a structurally distinct governance orientation.
4.3. Visibility as an Institutional Incentive Mechanism
A further question arises: why should making reasoning visible change organizational behavior? The answer lies not in any direct causal link between visibility and decision quality, but in how visibility alters institutional incentives.
When reasoning becomes institutionally reviewable, several changes occur in the decision environment. First, the anticipation of review alters the incentive structure of decision-making. Decision-makers who know their assumptions will be documented, their alternatives examined, and their treatment of dissent scrutinized face different incentives than those whose reasoning remains institutionally invisible. The mere prospect of reviewability—not the actual occurrence of review—can shift behavior. This is consistent with the extensive literature on how accountability pressures shape decision-making (Tetlock, 1992; Lerner & Tetlock, 1999).
Second, visibility enables ex-post learning. When assumptions are documented, their subsequent validation or falsification becomes an organizational resource. An organization that never records its assumptions loses the capacity to learn from their failure. An organization that systematically documents them creates the conditions for epistemic improvement over time. This is not about punishing flawed assumptions but about building institutional memory about the conditions under which assumptions prove reliable or unreliable.
Third, visibility redistributes the burden of justification. In its absence, those who challenge a decision must first reconstruct—often speculatively—the reasoning behind it before they can critique it. In its presence, the reasoning is already articulated; the challenger's task shifts from reconstruction to evaluation. This makes contestation less costly and more likely. Visibility does not guarantee that contestation will occur, but it reduces the institutional barriers to its occurrence.
Fourth, visibility creates reputational consequences for epistemic negligence. When reasoning is visible, those who consistently fail to examine assumptions, explore alternatives, or engage with dissent develop observable patterns that can inform personnel decisions, audit findings, and institutional accountability. The behavior itself may not change, but the institutional consequences of the behavior become more immediate. None of these mechanisms assumes that visibility guarantees better reasoning. They assume only that visibility changes the institutional conditions under which reasoning occurs, and that these changed conditions, on average and over time, are more conducive to robust deliberation than conditions of institutional invisibility.
4.4. Construct Hierarchy
To maintain conceptual clarity, the framework distinguishes four levels of abstraction:
· Level 1 — Governance Orientation: Epistemic accountability, a normative orientation toward making upstream deliberation institutionally legible and reviewable independent of outcome correctness.
· Level 2 — Evaluative Property: Deliberative visibility, an observable characteristic of decision processes: the extent to which assumptions, alternatives, and dissent are documented, exposed, and contestable—not the extent to which they are "good" but the extent to which they are present and reviewable in the institutional record.
· Level 3 — Institutional Mechanism: Decision process accountability, the specific mechanisms through which deliberative visibility is rendered reviewable within organizational settings.
· Level 4 — Operational Instruments: Traceability, contestability, and stress-testing, the concrete tools for implementing epistemic accountability.
4.5. Generative Mechanisms
The framework specifies five mechanisms through which deliberative visibility is expected to affect decision quality:
· Error detection mechanism: Structured dissent increases the probability that flawed assumptions will be identified before commitment (Janis, 1982).
· Blind spot reduction mechanism: Diverse perspectives reduce the risk that shared cognitive biases will go unchallenged (Tetlock, 2005).
· Hypothesis space expansion mechanism: Systematic alternative exploration counteracts satisficing and reduces the risk that superior options are never considered (Heuer, 1999; Simon, 1947).
· Assumption exposure mechanism: Explicit documentation transforms assumptions from cognitively costless background beliefs into propositions that invite scrutiny (Kahneman, 2011).
· Counterfactual activation mechanism: Prospective stress-testing counteracts overconfidence by forcing engagement with scenarios in which the preferred course of action fails (Klein, 1998).
These mechanisms are theoretically grounded and empirically testable. However, it bears repeating that they represent contingent empirical hypotheses, not foundational premises. The framework's central claim—that the recoverability of reasoning constitutes a missing dimension of accountability—does not depend on the empirical confirmation of these mechanisms. It depends on the observation that existing accountability architectures render reasoning insufficiently visible, and that this invisibility represents a correctable institutional deficit.
4.6. The Accountability Matrix: Existing vs. Epistemic Accountability
To clarify the framework's contribution, it is useful to contrast the questions asked by existing accountability mechanisms with those asked by epistemic accountability:
· Existing Accountability: Did rules get followed? Epistemic Accountability: Were assumptions exposed?
· Existing Accountability: Did outputs get delivered? Epistemic Accountability: Were alternatives explored?
· Existing Accountability: Was money spent properly? Epistemic Accountability: Was dissent surfaced?
· Existing Accountability: Were targets achieved? Epistemic Accountability: Was the reasoning process documented?
· Existing Accountability: Ex post evaluation of outcomes. Epistemic Accountability: Ex ante visibility of reasoning.
· Existing Accountability: Outcome accountability—what was achieved. Epistemic Accountability: Deliberative traceability—how was the decision thought through.
This matrix encapsulates the framework's core contribution: a dimension of accountability that existing mechanisms were never designed to address.
4.7. Positioning the Framework within Governance Theory
Epistemic accountability extends several established traditions while addressing a gap they leave unaddressed. It shares with deliberative governance (Habermas, 1996; Dryzek, 2000) a commitment to the epistemic value of structured deliberation, but differs in focus: deliberative governance is concerned with the forum—the public sphere, the legislative chamber—while epistemic accountability is concerned with the institutional apparatus that renders deliberation reviewable within administrative and executive settings. Deliberative governance evaluates the inclusiveness and legitimacy of discourse; epistemic accountability evaluates the recoverability and reviewability of reasoning. This is a sharper distinction than what has been drawn in earlier formulations of the framework, and it is essential to the framework's distinctive contribution.
It intersects with anticipatory governance (Guston, 2014; OECD, 2022) but focuses not on the content of scenarios considered but on the process through which they are generated and the institutional obligation to do so.
It resonates with evidence-based policymaking (Nutley, Walter, & Davies, 2007) but is agnostic about evidence type; its concern is whether the evidentiary basis for critical assumptions has been made explicit and subjected to scrutiny.
It draws on behavioral public administration (Grimmelikhuijsen et al., 2017; Moynihan, 2018) for its diagnostic vocabulary, but moves beyond diagnosis to institutional prescription: behavioral public administration has demonstrated that cognition matters; epistemic accountability proposes an institutional architecture for making cognitive processes reviewable.
It draws on procedural rationality (Simon, 1976) but makes procedural rationality accountable—not merely a cognitive practice internal to decision-makers but an institutional requirement subject to external review.
Epistemic accountability does not claim to supersede these traditions. It claims to integrate their insights into an accountability architecture that identifies a missing dimension—accountability for upstream deliberation—and proposes mechanisms for making that dimension governable.
---
5. Toward Operationalization: A Preliminary Architecture
This section outlines operational instruments corresponding to the three dimensions of deliberative visibility. They are illustrative starting points for empirical testing, not final products. To move from abstract ideals toward measurable constructs, each dimension is specified with observable indicators. A critical methodological clarification must be made: these indicators are not proposed as measures of reasoning quality—a construct whose validity would be deeply contested—but as measures of deliberative visibility, the extent to which the reasoning process has been made institutionally legible and recoverable. They are institutional observability indicators, not epistemic performance metrics. A decision process that documents many assumptions, explores many alternatives, and records dissent is not thereby guaranteed to be "better" reasoned; but it is more reviewable, and it is reviewability, not quality, that the framework seeks to operationalize. This distinction is important for the framework's methodological defensibility: the indicators reward documentation completeness, not reasoning excellence; they measure what was made visible, not what was thought.
5.1. Traceability
The first component is structured documentation of deliberation at the time decisions are made, requiring decision-makers to articulate the problem definition and its evidentiary basis, key assumptions and their vulnerability, alternatives considered and justification for selection, and anticipated risks and proposed mitigation. The purpose is to make explicit the reasoning that already occurs informally, rendering it available for subsequent review and institutional learning.
Observable indicators for assessing traceability—understood as indicators of documentation completeness, not reasoning quality—include:
· The number of explicit causal assumptions documented in the decision record,
· The presence or absence of identified "assumptions that if incorrect would fundamentally alter the decision,"
· The number of alternatives formally assessed with documented rationale for rejection,
· The existence of a risk anticipation section identifying plausible failure scenarios.
5.2. Contestability
The second component is a mechanism ensuring dissenting views are surfaced, documented, and addressed, requiring a formal opportunity for dissent, documentation of dissenting views, and substantive response from decision-makers. The existence of this mechanism is a structural feature of the decision process; its effectiveness is measured by accessibility and response quality, not dissent volume. Contestability transforms dissent from an informal, often suppressed, organizational dynamic into an institutionalized, documented dimension of the decision record.
Observable indicators for assessing contestability—understood as indicators of process accessibility, not dissent validity—include:
· The existence of a formal dissent recording mechanism in the decision process,
· The number of documented dissenting views in the decision record,
· The presence or absence of substantive responses to material dissent,
· Whether dissenting views can be submitted anonymously without career consequences.
5.3. Stress-Testing
The third component is a facilitated pre-mortem (Klein, 1998), where decision-makers imagine catastrophic failure and work backwards to identify plausible causes. The process surfaces implicit assumptions, overlooked risks, and unconsidered scenarios, directly counteracting overconfidence bias. Pre-mortem facilitation is proposed as a role for internal audit functions—a natural extension of their existing advisory and consulting role under frameworks such as the Global Internal Audit Standards (IIA, 2024).
Observable indicators for assessing stress-testing—understood as indicators of simulation thoroughness, not predictive accuracy—include:
· The presence of a documented pre-mortem or equivalent failure simulation in the decision record,
· The number and diversity of failure scenarios examined,
· Whether the pre-mortem generated new risks not previously identified in the risk register,
· Whether the results informed explicit modifications to the proposed course of action.
5.4. Integration
The three instruments are integrated through an auditable decision structure—a minimum procedural requirement satisfied before a strategic decision is finalized. It does not empower reviewers to block decisions. If a decision-maker proceeds despite identified risks, this is recorded as a documented override—conscious risk acceptance with stated justification. The documented override is not a failure of the framework; it is the framework functioning as designed, making visible the exercise of decision authority under conditions of acknowledged risk.
---
6. Non-Claims, Pathologies of Deliberation, Trade-Offs, Resistance, and Boundary Conditions
6.1. What This Framework Does Not Claim
The framework does not claim to guarantee good outcomes, eliminate politics, solve corruption, replace democratic legitimacy, eliminate uncertainty, or imply retrospective determinism. A decision reasonable ex ante remains reasonable even if it produced an adverse outcome. Most importantly, the framework does not claim to audit cognition directly; it audits the institutional record of deliberation—the assumptions documented, alternatives explored, and dissent recorded—not the internal mental processes of decision-makers.
6.2. Epistemic Failure vs. Political Rationality
Not all output-compliant but outcome-weak governance reflects epistemic failure. It is essential to distinguish between decisions whose outcomes are weak because reasoning was flawed or unexamined, and decisions whose outcomes are weak because political rationality dictated a suboptimal choice. In many cases, reasoning is adequate—risks are understood, alternatives are known, consequences are anticipated—but the decision is nonetheless taken in a direction that prioritizes coalition maintenance, rent distribution, symbolic politics, territorial bargaining, or electoral signaling over substantive outcome optimization.
This is not a failure of the framework; it is the framework properly bounded. Epistemic accountability does not prevent political choice or subordinate democratic judgment to technocratic optimization. It makes the trade-offs between epistemic robustness and political rationality visible and attributable. When a decision is taken for political reasons that diverge from the conclusions of a well-documented deliberative process—when risks are acknowledged and nonetheless accepted, when alternatives are explored and nonetheless rejected on grounds other than evidence—the documented override ensures that this choice is recorded, traceable, and available for subsequent democratic and institutional scrutiny.
The distinction between epistemic failure and political rationality is essential to the framework's credibility. It prevents epistemic accountability from being dismissed as naively technocratic or anti-democratic. It acknowledges that governance is irreducibly political, that optimization is not always the overriding value, and that democratic legitimacy may require decisions that are epistemically suboptimal. What epistemic accountability demands is not that politics be eliminated but that its exercise be made institutionally legible.
6.3. Power and the Limits of Epistemic Accountability
The framework must also contend with a more fundamental challenge: that reasoning visibility itself can become politically captured, and that the institutional conditions for contestability are shaped by structures of power that the framework, by itself, cannot overcome. This acknowledgement is not a concession of weakness but a necessary specification of the framework's boundary conditions.
Reasoning visibility does not operate on a level institutional field. In settings characterized by asymmetrical authority, bureaucratic domination, or epistemic hierarchy, the capacity to contest decisions—and to have that contestation recorded and taken seriously—is unequally distributed. Those with organizational power can shape what counts as "reasonable" deliberation, whose assumptions warrant examination, and which dissenting views deserve a response. The framework's contestability mechanisms, however well-designed, operate within these pre-existing power structures and can be captured by them.
Similarly, the definition of what constitutes adequate assumption documentation, sufficient alternative exploration, or genuine dissent is not politically neutral. Bureaucratic elites, professional experts, and dominant political actors may shape these standards in ways that exclude the perspectives of less powerful stakeholders while maintaining the formal appearance of deliberative robustness. The framework risks what could be termed technocratic exclusion—the use of apparently neutral procedural standards to delegitimize substantive disagreement that does not conform to elite epistemic norms.
These are genuine vulnerabilities, not peripheral concerns. They do not, however, undermine the framework's central claim. The appropriate comparison is not between epistemic accountability and some ideal institutional condition free of power asymmetries—no such condition exists. The appropriate comparison is between a governance architecture in which reasoning is at least potentially recoverable and contestable, and one in which it is systematically invisible. The former may be captured by power; the latter is already structured to serve it.
Moreover, visibility can, under certain conditions, serve as a resource for the relatively powerless. Documented assumptions can be challenged by those who were not in the room when they were adopted. Recorded dissent—even when overruled—can inform public scrutiny, legislative oversight, and future accountability processes. The framework does not eliminate power asymmetries, but it can create institutional footholds for contestation that would not otherwise exist.
Acknowledging these limitations strengthens rather than weakens the framework. It signals that epistemic accountability is not naively technocratic, that it understands the political conditions under which it would operate, and that its contribution is incremental and conditional rather than transformative and unconditional.
6.4. The Potential Pathologies of Deliberation
The framework does not assume that more deliberation is always better. Deliberation has its own pathologies, well-documented in organizational theory. The garbage can model of organizational choice (Cohen, March, & Olsen, 1972) demonstrates that decision processes can become arenas for problems, solutions, participants, and choices to interact in ways largely decoupled from rational problem-solving. Increased deliberation under these conditions may amplify noise rather than signal.
Bounded rationality overload occurs when the cognitive demands of structured deliberation exceed the information processing capacity of decision-makers, leading to simplification strategies that undermine the very robustness deliberation was meant to achieve. Decision latency—the prolongation of decision timelines through additional procedural requirements—can be costly in contexts where timeliness matters. Escalation complexity arises when multiple review layers generate coordination requirements that exceed organizational capacity. Incrementalism, as theorized by Lindblom (1959) in his analysis of "muddling through," suggests that many organizational decisions are necessarily incremental and adaptive rather than comprehensively rational, and that the attempt to impose comprehensive rationality on such processes may be counterproductive.
These are not arguments against epistemic accountability but specifications of the conditions under which it is most and least likely to be beneficial. The framework's boundary conditions explicitly exclude contexts where speed is paramount or where the decision stakes do not justify the procedural investment.
6.5. Governance Trade-Offs
Beyond the pathologies of deliberation itself, epistemic accountability involves genuine trade-offs with competing governance values:
· Speed-deliberation trade-off: Structured deliberation requires time. In emergency response contexts, speed is the overriding imperative and the opportunity cost of deliberation is unacceptably high. The framework is not designed for such contexts.
· Documentation-pragmatism trade-off: Any requirement for structured documentation risks being perceived—or becoming—an additional bureaucratic burden that displaces rather than enhances substantive deliberation.
· Contestability-consensus trade-off: Formal dissent mechanisms can generate performative rather than substantive challenge—dissent that is scripted, safe, and unlikely to disturb organizational hierarchies. In high power-distance cultures, dissent mechanisms may be captured by dominant actors to legitimize predetermined positions.
· Robustness-decisiveness trade-off: An overemphasis on deliberative visibility can lead to paralysis by analysis. The documented override mechanism is designed to mitigate this risk by preserving decision-maker authority.
· Transparency-vulnerability trade-off: The documentation of assumptions and dissent creates a record that can be used punitively in adversarial institutional environments. The context lock and safe harbor provisions discussed below are designed to address this concern, but the risk cannot be eliminated entirely.
6.6. Why Institutions Resist Epistemic Accountability
The framework faces predictable institutional resistance rooted in the political economy of accountability. Several mechanisms drive this resistance.
Blame avoidance is a well-documented feature of public sector behavior (Hood, 2011; Weaver, 1986). Epistemic accountability increases the traceability of reasoning, which means that flawed assumptions and overlooked alternatives become institutionally visible. For decision-makers operating in high-blame environments, this visibility constitutes a threat. The rational response may be to minimize documentation—to keep reasoning implicit, assumptions unstated, and dissent unrecorded—so that accountability for decisions remains diffuse rather than attributable.
Plausible deniability is a resource that epistemic accountability reduces. When reasoning is undocumented, decision-makers can, ex post, claim that risks were considered, alternatives were evaluated, or that the decision was forced by circumstances. Structured documentation eliminates this flexibility. Organizations that depend on strategic ambiguity to manage political relationships may resist frameworks that make ambiguity institutionally costly.
Coalition maintenance in political environments often requires that the reasoning behind decisions remain opaque to external scrutiny. Decisions that satisfy competing constituencies may depend on different justifications being presented to different audiences. Standardized reasoning documentation makes this strategic flexibility more difficult to sustain.
Symbolic compliance is the most likely form of institutional response to epistemic accountability mandates. Organizations may adopt the formal trappings—templates completed, forms signed, pre-mortems conducted—while decoupling these rituals from actual decision-making. This is precisely the dynamic Power (1997) identified in the audit society, and epistemic accountability is no more immune to it than any previous reform.
Acknowledging these resistance mechanisms serves two purposes. First, it signals that the framework does not assume institutional naivety. Second, it suggests that implementation design must anticipate these dynamics—for example, by building coalitions of institutional actors whose interests align with increased reasoning visibility, by starting with decisions where the benefits of traceability are highest, and by designing safe harbor provisions that reduce the personal risks of documentation.
6.7. The Tragedy of Recursive Ritualization
Every governance instrument relying on documentation is vulnerable to ritualization—the hollowing out of substantive practice while maintaining formal appearance. There is a tragic recursive quality to accountability reform: each new oversight layer, introduced to correct previous deficiencies, becomes subject to the same dynamics. Performance audit, risk governance, evidence-based policy, and anticipatory governance were all introduced as solutions to earlier limitations; each has generated its own rituals of compliance.
This is not merely an empirical observation but a theoretical implication of institutionalized legibility bias itself. Any accountability mechanism, by making certain organizational behaviors visible, incentivizes investment in precisely those visible behaviors. Over time, organizations learn to optimize for the metric rather than the underlying quality it was designed to proxy. The mechanism is self-undermining: accountability produces legibility, legibility produces gaming, gaming produces ritualization, ritualization produces demand for new accountability, and the cycle repeats. Each generation of reform becomes the problem the next generation must solve.
The framework is no more immune to recursive ritualization than any of its predecessors. Several safeguards are proposed: selective deployment on high-stakes decisions rather than universal application, independent facilitation by parties with professional credibility invested in process integrity, diagnostic monitoring of superficial compliance patterns—consistently unanimous decisions, uniformly low risk ratings, absence of documented dissent across multiple decisions—as triggers for deeper review, and explicit feedback loops comparing documented reasoning with actual outcomes. A framework that generates data about its own effectiveness—or ineffectiveness—is less susceptible to ritualization than one that does not.
The framework is intentionally minimalist, specifying procedural conditions necessary for recoverable reasoning rather than attempting exhaustive epistemic verification. This minimalism is itself a defense against recursive ritualization: fewer requirements mean fewer targets for gaming, and clearer requirements mean that gaming, when it occurs, is easier to detect.
6.8. Safe Harbor and Context Lock
The context lock principle assesses documented reasoning based on information reasonably available at the time, not on outcomes. A safe harbor provision establishes that decision-makers following the structured process have established a rebuttable presumption of due care. This provides no immunity for fraud, corruption, gross negligence, bad faith, or intentional concealment. The safe harbor applies exclusively to epistemic due process—the good-faith effort to document assumptions, explore alternatives, surface dissent, and stress-test reasoning. It does not shield corruption or intentional misconduct. By reducing the fear of ex-post punishment for reasonable uncertainty and good-faith documentation of risks that later materialize, the safe harbor reduces defensive bureaucracy and encourages the candid articulation of what is not known at the time of decision.
6.9. Boundary Conditions
The framework is most relevant for decisions characterized by high strategic significance, high uncertainty, high irreversibility, and multi-stakeholder complexity. It is unsuitable for routine operational decisions, emergency response, or highly standardized transactions.
6.10. Empirical Validation Agenda
The framework requires empirical validation. Priority questions include whether deliberative visibility can be proxied through observable indicators with sufficient reliability and validity, whether higher deliberative visibility correlates with better downstream outcomes controlling for sector, scale, and political context, whether decision-makers engage substantively or performatively with structured reasoning processes, and under what institutional conditions the framework is most effective.
A feasible initial study would code a sample of completed projects on the observable indicators specified in Section 5—assumption documentation completeness, dissent documentation, alternative exploration documentation, and stress-testing evidence—then test associations with cost overruns, delays, and utilization rates. The central falsifiable proposition is that projects exhibiting higher deliberative visibility—greater observable traceability of assumptions, alternatives, and dissent—are associated with lower rates of avoidable outcome failure, controlling for sector, scale, and political context. Randomized controlled trials on prospective decisions represent the strongest test of causality.
---
7. Conclusion
The historical trajectory of public accountability has been one of progressively expanding what institutions make visible: first expenditure, then compliance, then outputs, performance, and risk. Each expansion widened the evaluative reach of governance. Yet governance systems still largely treat reasoning itself as institutionally private—something that occurs within the minds of decision-makers, within the confines of meeting rooms, within the informal dynamics of bureaucratic hierarchy. The central argument of this paper is that contemporary governance failure increasingly arises not from the absence of oversight, but from an asymmetry between visible execution and invisible deliberation.
The concept of institutionalized legibility bias explains this asymmetry: organizations optimize for what accountability systems reward, and accountability systems reward what they can measure. The driving mechanism—auditable defensibility—ensures that organizations invest in the documentation of actions, the regularity of procedures, and the completeness of compliance records, while systematically underinvesting in the examination of assumptions, the exploration of alternatives, and the facilitation of dissent. The result is epistemic decoupling: a growing divergence between procedural accountability and substantive decision quality, an accountability architecture that is simultaneously more sophisticated and more hollow.
Epistemic accountability is proposed as a governance orientation that addresses this asymmetry. It is distinguished from procedural accountability not by advocating additional procedural requirements but by shifting the object of accountability itself—from the verification of actions to the reconstructability of reasoning. It is distinguished from deliberative governance by its focus on the institutional apparatus of reviewability rather than the democratic legitimacy of discourse. It distinguishes between epistemic failure and political rationality, acknowledging that many suboptimal decisions are not reasoning failures but political choices whose trade-offs should be made institutionally visible. It acknowledges that reasoning visibility operates within structures of power that can capture and shape it, and that its contribution is incremental and conditional rather than transformative.
Epistemic accountability does not audit cognition directly—it does not claim access to internal mental states—but instead audits the documentary record of deliberation: the assumptions documented, alternatives explored, and dissent recorded. It does not assume that more deliberation is always better—indeed it explicitly acknowledges the pathologies deliberation can produce. It does not assume immunity from the recursive dynamics of ritualization that have beset every previous accountability reform—instead, it builds diagnostic capacity for detecting its own ritualization into the framework itself. It acknowledges the political economy of resistance and the strategic reasons institutions have for keeping reasoning implicit.
Modern governance may never fully escape ritualization. The question is not whether accountability systems distort organizational behavior—they inevitably do—but whether they can be designed to distort it toward more epistemically recoverable forms of decision-making. The tragedy of recursive ritualization means that every accountability framework eventually becomes part of the problem it was designed to solve. The best that can be hoped for is a framework that, by making its own failure modes legible, creates the conditions for its own correction.
Epistemic accountability represents an attempt—not to guarantee correctness—but to close the asymmetry between visible execution and invisible deliberation by making the reasoning structures underlying consequential public decisions institutionally recoverable. The question for future governance, then, is no longer simply whether institutions can verify what government did, but whether institutions can reconstruct how government came to believe that doing it was justified in the first place.
Whether this orientation would reduce the incidence of output-compliant but outcome-weak governance is an empirical question. This paper has specified the generative mechanisms through which it might, the boundary conditions under which it is most likely to be effective, the trade-offs it entails, and the empirical strategies through which its claims can be tested and falsified.
The existing accountability architecture is incomplete. It governs the downstream consequences of decisions—financial integrity, procedural regularity, output achievement—but leaves the upstream origins of those decisions in an accountability vacuum. Closing that gap requires recognizing that accountability is not only about what was done and how, but about how it was thought through before it was done.
---
References
Bovens, M. (2007). Analysing and assessing accountability: A conceptual framework. European Law Journal, 13(4), 447–468.
Brunsson, N. (2002). The organization of hypocrisy: Talk, decisions and actions in organizations (2nd ed.). Copenhagen Business School Press.
Cohen, M. D., March, J. G., & Olsen, J. P. (1972). A garbage can model of organizational choice. Administrative Science Quarterly, 17(1), 1–25.
Dryzek, J. S. (2000). Deliberative democracy and beyond: Liberals, critics, contestations. Oxford University Press.
Espeland, W. N., & Sauder, M. (2007). Rankings and reactivity: How public measures recreate social worlds. American Journal of Sociology, 113(1), 1–40.
Fischer, F. (2003). Reframing public policy: Discursive politics and deliberative practices. Oxford University Press.
Flyvbjerg, B. (2017). Introduction: The iron law of megaproject management. In B. Flyvbjerg (Ed.), The Oxford handbook of megaproject management (pp. 1–18). Oxford University Press.
Goodhart, C. A. E. (1984). Monetary theory and practice: The UK experience. Macmillan.
Grimmelikhuijsen, S., Jilke, S., Olsen, A. L., & Tummers, L. (2017). Behavioral public administration: Combining insights from public administration and psychology. Public Administration Review, 77(1), 45–56.
Guston, D. H. (2014). Understanding "anticipatory governance." Social Studies of Science, 44(2), 218–242.
Habermas, J. (1996). Between facts and norms: Contributions to a discourse theory of law and democracy. MIT Press.
Heuer, R. J. (1999). Psychology of intelligence analysis. Center for the Study of Intelligence, Central Intelligence Agency.
Hood, C. (2011). The blame game: Spin, bureaucracy, and self-preservation in government. Princeton University Press.
Janis, I. L. (1982). Groupthink: Psychological studies of policy decisions and fiascoes (2nd ed.). Houghton Mifflin.
Kahneman, D. (2011). Thinking, fast and slow. Farrar, Straus and Giroux.
Kahneman, D., & Tversky, A. (1979). Prospect theory: An analysis of decision under risk. Econometrica, 47(2), 263–291.
Klein, G. (1998). Sources of power: How people make decisions. MIT Press.
Lerner, J. S., & Tetlock, P. E. (1999). Accounting for the effects of accountability. Psychological Bulletin, 125(2), 255–275.
Lindblom, C. E. (1959). The science of "muddling through." Public Administration Review, 19(2), 79–88.
March, J. G., & Simon, H. A. (1958). Organizations. Wiley.
Moynihan, D. P. (2018). A great schism approaching? Towards a micro and macro public administration. Journal of Behavioral Public Administration, 1(1).
Muller, J. Z. (2018). The tyranny of metrics. Princeton University Press.
Nutley, S. M., Walter, I., & Davies, H. T. O. (2007). Using evidence: How research can inform public services. Policy Press.
Organisation for Economic Co-operation and Development. (2022). Anticipatory innovation governance model in Finland. OECD Publishing.
Perrow, C. (1984). Normal accidents: Living with high-risk technologies. Basic Books.
Power, M. (1997). The audit society: Rituals of verification. Oxford University Press.
Scott, J. C. (1998). Seeing like a state: How certain schemes to improve the human condition have failed. Yale University Press.
Shore, C., & Wright, S. (2015). Audit culture revisited: Rankings, ratings, and the reassembling of society. Current Anthropology, 56(3), 421–444.
Simon, H. A. (1947). Administrative behavior: A study of decision-making processes in administrative organization. Macmillan.
Simon, H. A. (1976). Administrative behavior: A study of decision-making processes in administrative organization (3rd ed.). Free Press.
Strathern, M. (2000). The tyranny of transparency. British Educational Research Journal, 26(3), 309–321.
Tetlock, P. E. (1992). The impact of accountability on judgment and choice: Toward a social contingency model. Advances in Experimental Social Psychology, 25, 331–376.
Tetlock, P. E. (2005). Expert political judgment: How good is it? How can we know? Princeton University Press.
The Institute of Internal Auditors. (2024). Global internal audit standards. IIA Global.
Vaughan, D. (1996). The Challenger launch decision: Risky technology, culture, and deviance at NASA. University of Chicago Press.
Weaver, R. K. (1986). The politics of blame avoidance. Journal of Public Policy, 6(4), 371–398.