Beyond Maturity Scores: What Board Oversight of Technology Risk Actually Requires
Executive Summary:
Board oversight of technology risk remains anchored to maturity scores — metrics that describe how developed a program is, not whether critical services can continue operating under stress. As regulators across jurisdictions move toward requiring demonstrated resilience through tested performance evidence, the gap between what maturity reports convey and what boards are responsible for knowing has become a structural governance liability. The transition from maturity-based assurance to performance-based evidence is not a reporting upgrade; it is a reframing of what boards must demand in order to fulfil their fiduciary obligation to organizational continuity.
The Oversight Gap
Board oversight of technology risk has improved markedly over the past decade. Directors receive structured risk narratives, cybersecurity dashboards, and maturity assessments that offer a visible, comparative, and apparently objective account of where the organization stands. Regulatory attention has sharpened expectations further, and most large institutions can now point to governance frameworks, committee structures, and reporting cadences that would have been unusual fifteen years ago.
Yet across industries and jurisdictions, boards continue to struggle with a foundational question: whether their organization’s most critical services can continue to operate within acceptable tolerances during a severe but plausible disruption. The question is not new. What is becoming clear is that the reporting model most boards rely on is structurally incapable of answering it.
That reporting model is built on maturity scores. A maturity score describes how developed a program is: whether controls are documented, whether processes are repeatable, whether governance structures exist. These are meaningful inputs. But they do not describe how systems behave under stress, how dependencies fail, or whether recovery can occur at the speed and scale the enterprise requires. The gap between what maturity scores measure and what boards need to know is a structural limitation of the category.
This essay traces how that gap emerged, why it persists, and what the alternative looks like. The argument is narrower than abandonment: board oversight built exclusively on maturity-based assurance is no longer sufficient for the operational, interconnected, and time-compressed nature of modern technology risk. Boards need a different evidentiary standard, one grounded in demonstrated performance rather than documented intent.
I. How Maturity Models Became the Default
Maturity models did not arrive in boardrooms by accident. They emerged from software engineering, where the Capability Maturity Model developed at Carnegie Mellon in the late 1980s provided a structured way to evaluate how disciplined an organization’s software development processes were. The model offered a progression (from ad hoc and unpredictable at the lowest level to optimized and continuously improving at the highest) that gave leaders a legible account of where their teams stood and where they needed to invest. The logic was intuitive, the language was accessible, and the framework was portable. Within a decade, maturity models had been adapted into cybersecurity, information technology governance, and operational risk management.
The appeal to boards was immediate and understandable. Directors overseeing technology risk faced a translation problem that maturity models appeared to solve. Most board members lacked deep technical fluency. They needed a mechanism that could compress operational complexity into governance-legible metrics, something that allowed comparison across time, across business units, and against peers. A maturity score provided exactly that. A Chief Information Security Officer (CISO) could present a score of 2.8 this year compared to 2.4 the year prior, and a board could register trajectory without needing to understand the underlying controls, architectures, or dependencies that produced those numbers.
Frameworks reinforced this dynamic. The National Institute of Standards and Technology (NIST) Cybersecurity Framework, widely adopted across industries, introduced implementation tiers that functioned as maturity levels in practice. Audit firms developed maturity assessment methodologies that became standard deliverables. Industry benchmarking services offered comparative maturity data. Over time, the maturity score became the default unit of board-level technology risk reporting. Not because it was the most accurate measure of resilience, but because it was the most convenient translation layer between operational detail and governance judgment.
This trajectory deserves an important qualification. For organizations that are genuinely immature—those without baseline controls, without a governance framework, without a reporting discipline—a capability maturity model is the right starting tool. An organization that cannot articulate which of its services are critical, that has no documented processes for recovery, that lacks a structured approach to risk identification, is not in a position to produce performance evidence of resilience. That capability is simply beyond it. A maturity model gives such organizations a structured path to build from, a way to establish the foundational capabilities that performance-based oversight will eventually require.
The problem is that most organizations never graduate from maturity scores. What began as a developmental tool became a permanent reporting fixture. The convenience of the maturity score—its accessibility, its comparability, its resistance to ambiguity—made it self-reinforcing. Boards grew comfortable receiving maturity reports, CISOs grew comfortable producing them, and the incentive to move toward a more demanding evidentiary standard diminished with each reporting cycle. The starting tool became the only tool, and the governance model calcified around a metric that was never designed to answer the question boards most urgently need answered: not how mature the program looks, but whether the organization can survive a successful attack on its most critical services.
II. Where Maturity Scores Begin to Fail Boards
The simplicity of a maturity score is both its greatest strength and its fundamental limitation.
A board director does not need to understand cryptographic key management, network segmentation architectures, or identity federation protocols to know that 3.3 is higher than 3.0. That accessibility is what allowed cybersecurity to reach the board agenda in the first place. Before maturity scores, technology risk reporting was either too technical for governance audiences or too vague to be actionable. Maturity models provided a middle ground: a legible, trackable metric that gave boards a way to ask whether the trajectory was positive and whether investments were producing results. That contribution should not be dismissed.
But the same quality that makes maturity scores accessible also makes them misleading at the scale and speed of modern technology risk. A board that watches a maturity score progress from 3.0 to 3.3 to 3.7 over successive annual assessments experiences a sense of momentum. The trajectory appears positive. Resources are being deployed. The program is maturing. A ransomware event, a cloud provider outage, or a third-party compromise reveals that a 3.7 maturity score does not mean the institution can recover its critical business services within the tolerances the board assumed were in place. The score measured the existence and development of controls and processes. It did not measure whether those controls would hold under stress, whether recovery paths were viable, or whether dependencies had been mapped and tested.
This is a structural limitation inherent to the category, not a failure of any particular maturity model. Maturity models are designed to measure program completeness: whether processes are documented, whether roles are assigned, whether capabilities exist. They are not designed to measure operational performance: whether systems can recover within defined timeframes, whether dependencies fail gracefully, whether critical services continue when the assumptions embedded in their architecture prove wrong. The gap between these two questions is where board oversight most frequently breaks down.
Several specific failure modes illustrate the point. Maturity scores conflate activity with effectiveness. An organization can score highly on vulnerability management maturity while carrying critical unpatched exposures in production systems, because the maturity assessment evaluates the existence and repeatability of the scanning and remediation process, not the current state of the environment it is supposed to protect.
Maturity scores are static snapshots applied to dynamic risk. They capture program state at a point in time, but the risk environment they are meant to represent shifts continuously as new dependencies are introduced, cloud configurations change, and third-party providers update their own architectures. A score that was accurate in January may be misleading by March.
Maturity scores are assessed against documentation and self-reported evidence, not against disruption. An auditor evaluates whether a business continuity plan exists, whether it has been updated, and whether it follows a recognized structure. That assessment does not reveal whether the plan would actually work if invoked, whether the recovery sequence accounts for real dependency chains, or whether the teams responsible have rehearsed it under conditions that approximate a genuine disruption.
The structural point is this: maturity models measure the existence of controls and processes. They do not measure the behavior of systems and services under stress. For a board whose fiduciary responsibility includes ensuring that the organization can continue to operate through disruption, this distinction is the difference between overseeing a program and overseeing the outcome that program is supposed to produce.
III. Technology Risk as Behavioral and Operational
The maturity model’s inadequacy at the board level reflects a deeper misalignment: the way technology risk is categorized in most governance frameworks no longer matches the way technology risk actually behaves.
In most governance structures, technology risk (or its more specific variants, cybersecurity risk and information technology risk) is treated as a discrete risk type. It has its own committee, its own reporting line, its own assessment methodology, and its own maturity score. This structural separation made sense in an era when technology was a support function and its failure modes were largely contained within technical domains. A server outage affected a business unit. A security breach compromised a database. The blast radius was bounded, and the risk could be managed through dedicated controls and escalation procedures.
That era has passed. Technology is no longer a support function that operates beneath business processes. It is the medium through which business processes execute. Payment rails, customer onboarding, trading platforms, regulatory reporting, and internal communication all depend on digital infrastructure that is shared, outsourced, and tightly coupled. When a cloud service provider experiences a regional outage, the consequences are not confined to a technology domain. They cascade across every business function that depends on that provider’s infrastructure. When a third-party identity platform fails, the impact extends beyond access management to every service and workflow that relies on authentication through that platform. When a software supply chain is compromised, the exposure propagates through every organization that consumed the affected component.
In this environment, technology risk is embedded in the organization’s operational fabric: in the dependencies between systems, in the coupling between business processes and digital infrastructure, in the concentration of critical services on shared platforms, and in the assumptions that underpin architectural decisions. The Digital Operational Resilience Act (DORA), which became enforceable across the European Union in January 2025, reflects this recognition explicitly. DORA does not treat information and communication technology (ICT) risk as a subset of operational risk to be managed through a dedicated program. It treats ICT risk as operational risk, requiring management bodies to approve and oversee ICT risk management frameworks with the same governance rigor applied to credit, market, and liquidity risk.
This reframing has direct implications for board oversight. If technology risk is behavioral and operational, if it manifests through how systems interact, how dependencies cascade, and how architectural choices shape failure modes, then it cannot be overseen through documentation-based assurance alone. A maturity score describes the state of a program. It does not describe the behavior of a system. A board that receives an annual maturity report on cybersecurity is overseeing the program. A board that receives evidence of how critical services performed during the last resilience test, where dependencies failed, and what residual exposure remains is overseeing the risk. The distinction is not rhetorical. It determines whether the board has visibility into what would actually happen when disruption arrives.
IV. What Performance Evidence Actually Means
If maturity scores are structurally incapable of answering the question boards most need answered, the next question is immediate and practical: what should boards demand instead?
The answer is a different category of evidence altogether. Performance evidence is the demonstrated ability of an organization’s most critical services to continue operating within defined tolerances during a severe but plausible disruption. Each element of that definition carries specific meaning, and the precision matters. Boards that adopt the language of performance evidence without understanding what each term requires will reproduce the same oversight gap in a different vocabulary.
“Demonstrated” means tested, not documented. A business continuity plan that exists in a repository, has been reviewed annually, and follows a recognized template satisfies a documentation standard. It does not constitute performance evidence. Performance evidence requires that recovery capabilities have been exercised under conditions that approximate a genuine disruption: that systems have been failed over, that teams have executed recovery sequences, that dependencies have been stressed, and that the results have been recorded, analyzed, and reported to the board. The distinction between a plan that has been documented and a capability that has been tested is the distinction between intent and evidence.
“Critical services” means identified, not assumed. Most organizations have some notion of which services matter most, but few have completed the disciplined exercise of identifying which services, if disrupted beyond a defined threshold, would cause intolerable harm to customers, counterparties, markets, or the organization’s own viability. The United Kingdom’s Prudential Regulation Authority’s (PRA) concept of “important business services” reflects this requirement precisely: firms must identify the services that, if disrupted, would pose a risk to the firm’s safety and soundness, to policyholder protection, or to financial stability. Without this identification, performance evidence has no object. Testing becomes untethered from consequence.
“Defined tolerances” means the board has set explicit thresholds for acceptable disruption. An impact tolerance is not a recovery time objective established by a technology team. It is a board-level judgment about how much disruption the organization and its stakeholders can absorb before harm becomes intolerable. Setting impact tolerances requires the board to confront uncomfortable questions: how long can a payments service be unavailable before customers are materially harmed? How much data loss is acceptable in a trading environment before regulatory obligations are breached? These are governance decisions, not technical ones. They cannot be delegated to management without diluting the board’s oversight function.
“Severe but plausible” means the scenarios used to generate performance evidence must be realistic. This is a calibration challenge that boards should understand. Scenarios that are too mild produce false assurance: they confirm that recovery works under favorable conditions without revealing how it fails under stress. Scenarios that are implausibly extreme produce theatrical exercises that consume resources without generating actionable evidence. The standard established by the PRA and reinforced by DORA is severity calibrated to plausibility: scenarios that reflect disruptions the organization could credibly face, including the loss of a major third-party provider, a regional infrastructure outage, or a coordinated cyberattack targeting multiple systems simultaneously.
Performance evidence does not replace maturity models. Maturity assessments retain genuine value as program management tools, particularly for organizations that are still building foundational capabilities. An organization that lacks documented processes, assigned roles, or structured governance is not in a position to produce performance evidence. It first needs the programmatic foundations that maturity models help establish. Boards should not treat program maturity as a proxy for operational resilience once an organization has progressed beyond that foundational stage.
Performance evidence is also not a demand for continuous testing of everything. No organization can test every service, every dependency, and every failure mode on a continuous basis. The discipline requires prioritization: identifying which services are critical, which dependencies are most consequential, and which scenarios are most relevant, then testing those elements with sufficient rigor and frequency to produce evidence the board can rely on. The cadence of testing should align with the cadence of board reporting, so that the evidence directors receive reflects current operational reality rather than historical documentation.
The shift from maturity-based assurance to performance-based evidence is, at its core, a change in the question boards ask. The maturity question is: how developed is our program? The performance question is: can our most critical services continue within tolerance when disruption occurs? The first question can be answered with documentation and audit findings. The second can only be answered with tested, observed, and recorded evidence of how systems and teams actually behave under stress.
V. The Regulatory Imperative: Why Boards Cannot Wait
The shift from maturity-based assurance to performance-based evidence is becoming a regulatory requirement, not solely an intellectual argument about what constitutes good oversight. Across the jurisdictions that most directly shape board governance expectations, supervisors have independently arrived at the same conclusion: documentation-based assurance is insufficient, and boards must be able to demonstrate that critical services can continue operating under stress. The convergence is not coincidental. It reflects a shared recognition that the nature of technology risk has outpaced the oversight models that most boards employ.
DORA provides the most explicit articulation of this expectation. The regulation, which became enforceable in January 2025, does not treat ICT risk as a peripheral concern to be managed through a dedicated compliance program. It treats ICT risk as a core operational risk that requires direct management body oversight. Articles 5 and 6 of the regulation establish that management bodies must approve and actively oversee the ICT risk management framework, bear ultimate responsibility for managing ICT risk, and define roles and responsibilities for all ICT-related functions. The language is not aspirational. It creates a governance obligation that supervisory authorities can examine and enforce.
DORA’s resilience testing requirements reinforce the evidentiary standard. Financial entities classified as significant must conduct threat-led penetration testing at least every three years, using scenarios that reflect the current threat landscape. The regulation does not accept testing against documentation or self-assessment as sufficient. It requires demonstrated performance under realistic conditions. For boards of DORA-regulated institutions, the implication is direct: supervisors will evaluate not only whether an ICT risk management framework exists, but whether the organization has tested its ability to operate through disruption and whether the board has overseen that process.
The United Kingdom’s approach, while structurally distinct from DORA, arrives at a parallel conclusion through a different regulatory architecture. The PRA’s supervisory statements SS1/21 and SS2/21 established the operational resilience framework that requires firms to identify their important business services, set impact tolerances for those services, and demonstrate that they can remain within tolerance during a severe but plausible disruption. The framework moved through a transitional period during which firms were expected to identify services and set tolerances. That phase is complete. Firms are now in the demonstration phase. The expected output is proof. The PRA’s framework is notable for its precision: it defines what must be identified, what thresholds must be set, and what evidence must be produced. Boards that cannot articulate which services are important, what tolerances have been defined, and whether testing has confirmed the organization’s ability to remain within those tolerances are exposed to supervisory challenge.
In the United States, the regulatory path has been less unified but no less consequential. The United States Securities and Exchange Commission’s (SEC) amendments to Regulation S-K, specifically Item 106, require public companies to describe the board’s oversight of risks from cybersecurity threats and to disclose management’s role in assessing and managing those risks. This disclosure obligation, which took effect for fiscal years ending on or after December 15, 2023, transforms board oversight of cybersecurity from an internal governance matter into a public record. Investors, regulators, and litigators can now evaluate what a board claims to oversee and compare that claim against the organization’s actual performance during incidents disclosed under the companion 8-K reporting requirement. A board that discloses robust cybersecurity oversight in its annual filing but cannot produce evidence of what it actually oversees faces reputational, legal, and regulatory exposure.
The Basel Committee on Banking Supervision’s (BCBS) principles for operational resilience, published jointly with the Financial Stability Board (FSB), have reinforced these jurisdictional developments at the international level. The FSB’s framing, which emphasizes the continuity of critical functions rather than the solvency of individual institutions, has shaped the vocabulary and priorities that national regulators have subsequently codified or enforced. The National Association of Corporate Directors’ (NACD) 2026 Director’s Handbook on Cyber-Risk Oversight has translated this regulatory trajectory into governance practice, establishing six principles that frame board-level cybersecurity oversight as a core governance obligation rather than a technical delegation.
The regulatory convergence across these jurisdictions carries a specific message for boards that still rely primarily on maturity-based reporting. Each of these frameworks, through different mechanisms and different legal structures, points in the same direction: toward demonstrated resilience, tested under realistic conditions, overseen by governance bodies that understand what critical services are at stake and what tolerances have been set. Maturity scores may continue to serve as useful program management inputs, but they do not satisfy the evidentiary standard that regulators are now establishing. Boards that wait for their own regulators to issue prescriptive guidance before making this transition may find that the transition is made for them, under less favorable conditions.
VI. The Executive as Translator
The case for performance evidence is structural, regulatory, and operationally necessary. The immediate practical question is how that evidence reaches the boardroom. The answer depends almost entirely on the CISO, the Chief Information Officer (CIO), and the Chief Risk Officer (CRO), because these executives occupy the only position in the organization that bridges operational detail and governance judgment. The translation function they perform is the critical bottleneck in the shift from maturity-based to performance-based oversight.
Current CISO-to-board reporting is built around the maturity model for a reason that has nothing to do with the maturity model’s analytic merits. It is built around the maturity model because the maturity model is the established translation layer. A CISO who presents a maturity score is performing a compression: taking the full complexity of a cybersecurity program, its controls, its gaps, its dependencies, its recovery capabilities, its residual exposure, and rendering it as a number that a non-technical board can receive, compare, and track over time. That compression solves a genuine governance problem. Directors who oversee dozens of risk domains cannot absorb operational detail in each one. They depend on executives who can translate complexity into judgment.
The difficulty is that the compression maturity scores provide has become a substitution. The board does not receive a simplified representation of operational reality. It receives a different category of information altogether: a description of program development rather than a description of system behavior. The CISO who reports that vulnerability management maturity has improved from 3.1 to 3.4 is describing the program. The CISO who reports that the last resilience test revealed the payments platform recovered within tolerance but the customer onboarding service did not, that the root cause was an unmapped dependency on a shared authentication service, and that remediation is underway with a target completion date is describing the organization’s actual operational posture. The second report requires the board to engage with more specificity, but it answers the question the board is actually responsible for overseeing.
Shifting to performance-based reporting requires CISOs to hold three bodies of knowledge simultaneously: (1) architectural visibility into how critical services are constructed and where their dependencies lie, (2) testing data that reveals how those services and dependencies behave under stress, and (3) tolerance definitions that establish the thresholds against which performance is measured. This is not a communication skill that can be developed through presentation coaching or board-readiness training. It is a structural capability that depends on whether the organization has identified its critical services, mapped their dependencies, tested their continuity, and defined the tolerances against which performance is judged. A CISO who lacks any of these inputs cannot translate what the organization does not yet know about itself.
The incentive structure compounds the challenge. Boards that are accustomed to maturity reporting reward CISOs who deliver it fluently. A polished maturity presentation, complete with trend lines, peer benchmarks, and color-coded risk ratings, satisfies the reporting expectation without forcing the board to confront uncomfortable operational realities. A performance-based report, by contrast, may reveal that critical services cannot recover within tolerance, that dependencies are poorly understood, or that testing has exposed gaps the organization has not yet remediated. Delivering that report requires the CISO to present findings that may generate difficult questions, to acknowledge residual exposure the organization has accepted, and to do so in language that is precise enough for governance without being so technical that it loses the audience.
This is why the translation challenge is structural, not personal. The shift from maturity-based to performance-based board reporting cannot happen through CISO initiative alone. It requires boards to change what they ask for, organizations to build the operational infrastructure that generates performance evidence, and executives to develop the translation capability that renders that evidence in governance-legible form. The NACD’s 2026 Director’s Handbook recognizes this dynamic, framing board-level cybersecurity oversight as a governance obligation that requires directors to understand the difference between program maturity and operational resilience. The executive who can bridge that gap is more than a better communicator. That executive is performing a governance function that the current reporting model does not require, but that the emerging regulatory environment will demand.
VII. Oversight That Scales with Complexity
The most common objection to performance-based oversight is that it demands too much of boards. Directors oversee dozens of risk domains, meet several times a year at most on risk matters, and lack the technical background to evaluate service-level recovery data. If maturity scores succeeded because they compressed complexity into governance-legible metrics, the argument runs, then replacing them with service-specific performance evidence will overwhelm the oversight function rather than strengthen it.
The objection deserves a direct response because it reflects a genuine governance constraint. Boards are not operational bodies. They cannot and should not evaluate every resilience test, every dependency map, or every tolerance breach. The complexity objection does not argue against performance-based oversight. It argues for structuring it properly.
Performance evidence does not require boards to absorb more information. It requires them to absorb different information, organized around a smaller number of higher-consequence questions. Three structural mechanisms make this feasible.
The first is tiered reporting. Not every service in an organization requires board-level attention. Critical services, those whose disruption would cause intolerable harm to customers, counterparties, markets, or the organization’s viability, do. Management retains oversight of the broader service portfolio. The board focuses on whether the organization’s most consequential services can continue within tolerance. This is a narrower oversight aperture than most maturity-based reporting provides, where the board reviews an aggregate score that blends critical and non-critical capabilities into a single metric. Tiered reporting is more focused, not more complex.
The second is exception-based escalation. Maturity-based reporting presents the board with the full picture: scores across domains, trend lines, peer comparisons. Performance-based reporting inverts this logic. The board receives confirmation that critical services remain within tolerance, with detailed reporting triggered only when tolerances are breached or when testing reveals that continuity cannot be maintained. This model mirrors how boards already oversee other risk domains. Credit committees do not review every loan. They review exceptions to risk appetite. Audit committees do not re-examine every control. They review material weaknesses. Applying the same escalation logic to technology resilience aligns it with established governance patterns rather than creating a new and unfamiliar burden.
The third is cadence alignment. If the organization tests the resilience of its critical services on an annual cycle, board reporting on those services should align with that cycle so that directors receive evidence that reflects current operational reality. Maturity assessments, by contrast, often follow their own reporting calendar, disconnected from when the organization last tested its actual capabilities. The result is a board that reviews maturity data generated months earlier, applied to a risk environment that has changed since the assessment was completed. Aligning board reporting cadence with testing cadence ensures that the evidence directors receive is both current and grounded in observed behavior.
These mechanisms are not speculative. They reflect the governance architecture that DORA and the PRA’s operational resilience framework already imply. DORA requires management bodies to oversee ICT risk with the same rigor applied to credit and liquidity risk. The PRA requires firms to demonstrate that important business services can remain within impact tolerances. Both frameworks assume that boards will focus on the services that matter most, review evidence of how those services perform, and escalate attention when performance falls below defined thresholds. The structural infrastructure for performance-based oversight already exists in regulatory expectation. What remains is for organizations to build the operational capability to deliver it.
The complexity objection also fails to account for a dynamic examined in AI Increases the Need for Resilience, Not Automation Alone: the accelerating role of artificial intelligence in compressing decision timescales and operationalizing governance weaknesses. As AI-driven automation extends into security operations, incident response, and infrastructure management, the gap between what a maturity score describes and what the organization is actually doing widens further. A maturity assessment conducted annually cannot capture the pace at which AI-enabled systems introduce new dependencies, alter operational behavior, or shift the organization’s risk profile. Performance evidence, grounded in tested behavior rather than documented intent, is more responsive to this acceleration precisely because it measures what the organization can do, not what its documentation says it should be able to do.
Maturity models retain value in this structure. They serve as the program management layer that ensures foundational capabilities are in place, that governance processes function, that roles and responsibilities are defined. Maturity information should still occupy the management layer of oversight, informing CISOs, CIOs, and operational risk leaders about program health, while the board layer focuses on whether the organization’s critical services can actually continue through disruption. This clarifies where maturity models belong in the oversight architecture.
VIII. Confidence Comes from Performance, Not Apparent Maturity
Maturity models entered boardrooms to solve a legitimate governance problem: the need to compress technical complexity into a metric that non-technical directors could receive, compare, and act on. They succeeded in that narrow function, and for organizations still building foundational capabilities, they remain the right tool. The problem is that many large institutions long ago passed the developmental stage where maturity scoring provides its greatest value, yet the reporting model delivered to their boards has rarely evolved to reflect that progression. The metric designed to help organizations build programs became the metric boards use to judge resilience, and those are fundamentally different questions.
The distinction between program maturity and operational performance is a structural gap with material consequences. A board that receives a rising maturity score develops confidence in the organization’s resilience trajectory. That confidence is grounded in a category of evidence that describes how developed a program is: whether controls are documented, whether processes are repeatable, whether governance structures exist. None of these inputs, valuable as they are for program management, tell the board whether a ransomware attack would be contained before it reaches critical systems, whether a cloud provider outage would leave core customer services unavailable beyond acceptable limits, or whether recovery can occur at the speed the enterprise’s regulatory obligations and customer commitments require. The confidence the board holds is real. The evidence that produced it does not answer the question the board is responsible for.
The argument has been deliberate in what it does not claim. It does not argue that governance is insufficient, that documentation is unnecessary, or that maturity models should be discarded. A governance framework that defines accountability, establishes risk appetite, and creates reporting disciplines is a prerequisite for performance-based oversight. Documentation of processes, controls, and recovery plans is essential operational infrastructure. Maturity models that help organizations build structured capabilities from an undisciplined baseline serve a function that no other tool serves as well. These tools serve a different purpose than the one boards have increasingly asked them to serve. They describe intent, structure, and program development. They do not describe behavior under stress.
Boards across sectors and jurisdictions consistently report they cannot answer a foundational question: whether their organization’s most critical services can continue to operate within acceptable tolerances during a severe but plausible disruption. That question, stated plainly, is the question boards exist to ask. Fiduciary responsibility for organizational continuity is not delegable to management, to technology teams, or to audit functions. It belongs to the board. The reporting model most boards depend on is structurally incapable of providing the answer. Not because the model is flawed, but because it was designed to answer a different question: how mature is the program? The maturity answer is useful. It is not the answer the board needs.
Performance evidence closes this gap by changing the unit of analysis. The unit is no longer the control, the process, or the program. It is the critical service and its demonstrated continuity under stress. A board that receives performance evidence knows which services have been tested, where tolerances have been met, where they have been breached, what dependencies failed, and what residual exposure remains. That evidence does not eliminate risk. It does not guarantee continuity. What it provides is something maturity scores structurally cannot: a factual account of how the organization actually performs when its assumptions are challenged. The difference between apparent maturity and demonstrated performance is, in the end, the difference between oversight based on belief and oversight based on evidence.
Conclusion
The regulatory trajectory, the operational reality, and the fiduciary obligation all converge on the same point. Boards that oversee technology risk through maturity-based assurance are governing a representation of the organization, not the organization itself. The representation may be favorable. The maturity score may be rising. The governance structures may be well documented. Yet none of this answers the question that the PRA, DORA, and the SEC’s disclosure requirements have now placed squarely before directors: can this institution continue to deliver its most critical services when disruption arrives?
That question will not recede. The interdependencies that define modern technology infrastructure will grow more complex, not less. The concentration of critical services on shared platforms will deepen. The pace at which AI compresses decision timescales will accelerate. Each of these dynamics widens the gap between what a maturity score describes and what the organization actually faces. Boards that close this gap proactively, by demanding performance evidence, structuring oversight around critical services, and holding management accountable for tested continuity rather than documented intent, will find that the transition strengthens their governance position rather than complicating it. Boards that wait may find the gap closed for them, by regulators whose expectations have already moved, or by disruptions that expose what maturity narratives concealed.
The distinction between apparent maturity and demonstrated performance is durable because it is structural, not circumstantial. Programs will continue to mature. Documentation will continue to improve. Governance frameworks will continue to evolve. The question is whether the evidence boards receive reflects what the organization can actually do when it matters most. Confidence grounded in tested performance is the right standard. It treats resilience as a capability to be proven, not a score to be reported.
Oritse J. Uku is a CISO and resilience architect with two decades across military intelligence, global financial services, and enterprise cybersecurity. He authored the Resilience Operating Model (ROM), a framework for building institutions that perform under stress, govern risk at scale, and demonstrate regulatory readiness.
For interviews, briefings, or commentary on cybersecurity, architecture, or operational resilience, please visit the Media & Speaking page.
Discover more from Oritse J. Uku
Subscribe to get the latest essays sent to your email.