Explainability as an Architectural First-Class Concern
Introduction
We stand at a peculiar inflection point in financial history. The algorithms making credit decisions, fraud determinations, and trading executions increasingly operate beyond the comprehension of those who deploy them. This is not a failure of engineering. It is a failure of architectural philosophy, particularly in the context of AI explainability.
The global banking industry has embraced artificial intelligence with the fervor of converts, yet it has done so while importing a fundamental tension: the opacity inherent in modern machine learning systems collides directly with the regulatory imperative for AI explainability. The Basel Committee’s recent consultative papers on AI governance, the EU AI Act’s high-risk classifications, and the Federal Reserve’s model risk management guidance all point toward an uncomfortable truth. Our current architectures were not designed for algorithmic accountability. This assessment examines the structural dimensions of this challenge—the technical architecture required for explainability, the leadership disposition required for sustainable AI governance, and the strategic implications for institutions that fail to act.
Problem Description and Importance to Organizational Strategy
Consider the irony of our position. Banks operate within perhaps the most scrutinized regulatory environment on Earth. They maintain exhaustive audit trails for every transaction. They document every policy exception. They stress-test portfolios against scenarios that would make dystopian novelists blush. Yet they deploy neural networks that cannot articulate why they denied a mortgage application or flagged a transaction as potentially fraudulent.
This is not merely a compliance risk. It is an existential architectural flaw. As Jonathan Knee argues compellingly in The Platform Delusion, competitive advantages built on technological opacity rather than genuine structural moats inevitably erode (Knee, 2021). Banks deploying AI without interpretability infrastructure are constructing competitive positions on sand.
The technical debt accumulating in our AI systems differs qualitatively from traditional code debt. When a legacy COBOL system becomes unmaintainable, we understand its behavior even if we cannot easily modify it. When a deep learning model drifts or exhibits emergent bias, we may not even recognize the degradation until regulatory examination or customer litigation reveals it. The distinction is critical. Traditional technical debt is a known liability on the balance sheet of system maintenance. AI technical debt is an unknown liability, compounding silently, with the potential to surface as regulatory action, reputational damage, or systemic risk.
Initially, based on an assessment of AI deployments across major financial institutions, the following issues were found:
- No working governance relationship exists between the data science teams building models and the compliance organizations accountable for their outputs. Model development is headquartered in innovation labs and technology centers; regulatory accountability sits in risk management and legal. Both have board-level representation, yet neither has architectural authority over the other.
- Model deployment velocity is escalating, interpretability infrastructure is non-existent, and regulatory exposure is mounting.
- Senior technology leadership, in many institutions, appears to be adopting AI as a branding exercise, doing almost nothing to address the structural requirements for algorithmic accountability.
- AI systems, of which there are dozens across credit, fraud, trading, and customer segmentation, all operate with excess opacity and are running without explainability instrumentation. AI systems are capital and talent intensive, and to recover investment they need to operate with the trust of regulators and customers alike.
- Both training data pipelines and model serving infrastructure were built locally within business units. Potential for significant enterprise-wide governance capabilities is lost as a result of no centralized AI architecture.
- Inference audit logging and feature lineage tracking are redundant or absent across each business line.
- Model risk management treats a once-a-year regulatory filing the same as an ongoing production credit decisioning system.
- State of the art interpretability techniques—SHAP values, counterfactual explanations, attention visualization—are not leveraged to build trust in the systems or grow institutional confidence.
- Corporate communications regarding AI governance are non-existent to regulators, customers, and internal stakeholders.
- Stakeholders are at a breaking point. Within five years, unexplainable AI systems will be as impermissible in banking as undocumented capital calculations are today.
The Architectural Perspective: Designing for Explainability
The solution demands we reconceptualize AI not as a capability to be bolted onto existing systems, but as a distributed architectural concern requiring its own governance plane. Drawing from Brendan Burns’ insights in Designing Distributed Systems, we must treat model interpretability the way we treat observability in microservices—as a non-negotiable infrastructure component rather than an afterthought (Burns, 2018).
Concretely, this means architecting for what I call “algorithmic provenance”—the ability to reconstruct, at any future point, the exact reasoning chain that produced a given decision. The scope of work required to accomplish this transformation necessitates a new architectural structure with strong coupling between model development, governance, and operations. The following capabilities must be established:
Immutable Feature Stores. These must capture not just the features used in model inference, but the lineage of those features back to source systems. When a regulator asks why a customer was classified as high-risk eighteen months ago, we must be able to reconstruct the precise data state that existed at inference time. This is the AI equivalent of maintaining exhaustive audit trails for every transaction—a discipline banks already practice in every other domain.
Model Versioning Infrastructure. Trained models must be treated as immutable artifacts, analogous to container images, with cryptographic attestation of their training data, hyperparameters, and validation metrics. The disposition of model versions is critical—what to do with deprecated models, their training artifacts, validation records, and production histories. Governance recovered from proper versioning practices is significant.
Inference Audit Logging. This must capture not just inputs and outputs, but intermediate attention weights, feature importances, and uncertainty estimates. This is computationally expensive. It is also non-negotiable. Project monitoring of all inference events must alert governance teams to any model behaviors more than 15% off target for accuracy, fairness, or stability.
Counterfactual Explanation Services. These must generate, on demand, the minimal input perturbations that would have changed a decision—answering the inevitable question, “What would the customer have needed to do differently?” Communication programs for each stakeholder category—regulators, customers, internal risk committees—must be developed to translate these counterfactuals into language appropriate to each audience.
Sam Newman’s principles in Building Microservices apply here with particular force: these capabilities must be implemented as autonomous, independently deployable services with well-defined contracts, not as monolithic additions to existing model serving infrastructure (Newman, 2015). The organization required to support this architecture is not the traditional transactional structure organized by business unit. It is a matrix structure, with governance as a cross-cutting concern that intersects every business line deploying AI.
The Leadership Perspective: From Technological Enthusiasm to Philosophical Rigor
Technical architecture alone cannot solve this challenge. The deeper problem is cultural—a leadership disposition that has confused AI adoption with AI maturity.
Simon Sinek’s concept of the “infinite game” resonates here (Sinek, 2019). Banks pursuing AI as a finite game—a race to deploy more models faster than competitors—will inevitably sacrifice the interpretability infrastructure that enables long-term sustainability. The institutions that will thrive are those whose leaders understand AI governance as an infinite game, where the objective is not to “win” but to remain worthy of continued participation in the financial system.
Understanding AI Explainability in Banking
This requires a fundamental shift in how we measure AI program success. Metrics focused solely on model accuracy, inference latency, and deployment velocity incentivize precisely the opacity that will eventually become our undoing. We must elevate explainability coverage, audit trail completeness, and regulatory examination readiness to equal standing in our AI scorecards. The measurement framework must be specific in nature: if these performance indicators drift, do they still track within predefined acceptable levels of governance maturity?
Peter Block’s work on stewardship offers additional guidance. Block argues that genuine accountability emerges not from compliance mechanisms but from a felt sense of ownership for outcomes (Block, 2013). Our data scientists and ML engineers must internalize that they are not merely building prediction systems. They are making decisions that affect human lives—decisions for which they bear moral responsibility regardless of whether the algorithm can articulate its reasoning.
This is uncomfortable territory for technologists trained to view their work as objective and value neutral. It is also the truth. The architect who designs an unexplainable credit scoring system is complicit in every unjust denial that system produces, just as surely as the loan officer who personally rejected the application. The management team must be upgraded—not in the sense of replacement, but in the sense of expanding their conception of what it means to build and govern intelligent systems. Make every effort to upgrade the skills of the team from the boardroom to the data lab, to ensure that all are working as a team with the capabilities required to perform the required tasks and activities.
A Call to Architectural Conscience
The regulatory tide is turning. Within five years, I predict that unexplainable AI systems will be as impermissible in banking as undocumented capital calculations are today. The institutions that move now to architect for algorithmic accountability will find themselves with sustainable competitive advantages—not because their models are more accurate, but because their models can be trusted, audited, and defended.
The question I leave with my fellow architects and technology leaders is this: When the next financial crisis occurs—and it will occur—and when the post-mortem reveals that algorithmic systems amplified contagion or concentrated risk in ways their deployers did not understand, will your architecture be part of the problem or part of the solution?
So many lessons remain to be learned in this domain. The competitor who acquires AI capability without merging it into a governance-ready architecture—who instead merges the new technology into an obsolete operating model—will find that the acquisition was a disaster. The institutions that thrive will be those that recognized AI governance not as a constraint on innovation but as the structural foundation that makes innovation sustainable.
We have a narrow window to choose wisely. The philosopher-architect in me believes this is not merely a technical challenge but a moral one. The systems we build encode our values. Let’s ensure they encode values we are prepared to defend.
References
Block, P. (2013). Stewardship: Choosing Service Over Self-Interest. San Francisco, CA: Berrett-Koehler Publishers.
Burns, B. (2018). Designing Distributed Systems: Patterns and Paradigms for Scalable, Reliable Services. Sebastopol, CA: O’Reilly Media.
Knee, J. (2021). The Platform Delusion: Who Wins and Who Loses in the Age of Tech Titans. New York, NY: Portfolio/Penguin.
Newman, S. (2015). Building Microservices: Designing Fine-Grained Systems. Sebastopol, CA: O’Reilly Media.
Sinek, S. (2019). The Infinite Game. New York, NY: Portfolio/Penguin.


