Skip to main content
Lab Notes
AI Governance

AI Governance in 2026: What's Actually Working in Saudi Arabia and the Gulf

Nora Al-Rashidi|March 1, 2026|10 min read

There is a particular kind of governance document that gets produced in organizations across Saudi Arabia: carefully worded, comprehensively structured, aligned to the right international standards, and consulted by almost nobody. It satisfies auditors. It does not satisfy the operations manager whose AI routing system misrouted three refrigerated trucks on a Tuesday. It does not satisfy the loan officer who received an AI-generated credit recommendation with no way to understand the reasoning. It does not satisfy the regulator who, on inspection, discovers that the policies describe a governance function that exists on paper and nowhere else.

Across Saudi government agencies, financial institutions, logistics companies, and healthcare organizations, a pattern separates the organizations building AI governance that works from those building governance that merely looks like it should work. The distinction is not technical capability, budget, or the sophistication of the frameworks being adopted. It is whether governance has been designed to be used — by the people who operate AI systems every day — or designed to be filed.

The Context Has Shifted

Through 2024, AI governance in Saudi Arabia was, for most organizations, a compliance exercise with loose consequences. The National AI Ethics Principles existed. ISO 42001 was on the horizon. SDAIA's enforcement posture was visible mainly through principle-publishing rather than active scrutiny. Organizations could maintain governance documentation of variable quality without facing immediate pressure to demonstrate that it functioned.

That has changed. SDAIA has moved from publishing principles to running active certification programs. The NCA has begun requiring documented AI governance from suppliers in regulated food and logistics sectors, creating procurement pressure that reaches organizations well beyond those directly regulated. SAMA has issued AI guidance with real audit implications for the financial services sector. Saudi Arabia's international partners in defense, infrastructure, and health are increasingly treating AI governance as a due diligence criterion in contract evaluations. The organizations that built governance capacity before this shift are positioned differently from those that did not, and the cost of closing the gap under deadline is real.

The Sequence Error

The most consistent failure pattern in Saudi AI governance is not negligence but sequence error. Organizations begin with philosophy: drafting elaborate AI ethics frameworks, holding stakeholder workshops, commissioning principle documents, running all-staff training on responsible AI. The documents produced are often well-written and substantively thoughtful. And yet when an AI system fails — when a routing algorithm produces nonsensical outputs, when a credit scoring model behaves erratically on a population segment it has not encountered before, when a clinical AI tool generates an output no clinician can explain — the governance apparatus offers no operational response. The principles are on the wall. Nobody knows what to do.

The organizations making visible progress have inverted this sequence. They begin with incidents and incident response — not hypothetical incidents, but the actual failure modes that have occurred or that the system's operators can identify as likely. An organization deploying an AI routing system asks, before anything else: what happens when this system routes a high-value cold-chain shipment incorrectly? Who is notified? What is the manual fallback? What is the investigation procedure? The documented answers constitute the first meaningful layer of AI governance, because they are operational infrastructure, not aspiration.

This sequencing principle has a compounding effect. Organizations that establish incident response capability first build, as a side effect, the monitoring infrastructure required to detect incidents. That monitoring infrastructure then enables the transparency that more sophisticated governance layers require. Principle-first governance produces documents that gather dust; incident-first governance produces infrastructure that gets used.

Human Oversight That Works

A common governance overcorrection, particularly in organizations responding to regulatory pressure for the first time, is to require human approval for every AI decision. The reasoning is understandable: if AI can fail unpredictably, human review provides a safety net. The operational consequence is that the safety net destroys the value the AI was deployed to create, and produces worse oversight in the process.

When every decision requires approval, reviewers become rubber stamps. The volume of approvals exceeds any individual's genuine capacity for evaluation. The oversight becomes nominal — documented, but not substantive. Regulators who understand AI operations recognize this pattern and do not treat it as meaningful governance.

What distinguishes effective human oversight is precision about which decisions warrant it. In a logistics context, that means shipments above a certain value threshold, routes through high-temperature zones, movements to destinations the system has not been validated on. In a financial context: loan decisions above a defined amount, credit reassessments for customers in protected categories, outputs that feed directly into a customer communication. In a healthcare context: AI-generated recommendations that contradict the clinical judgment of the responsible physician.

The underlying question is not whether to review a decision, but what the worst-case consequence is if the AI is wrong. Where the answer involves significant financial, safety, reputational, or regulatory exposure, that decision point warrants human review. Where it does not, automation can proceed with continuous statistical monitoring. Organizations that have done this analysis rigorously find that meaningful oversight can be concentrated on a small fraction of decision types. The oversight becomes more effective because it is focused rather than diffuse.

Transparency First

Before approval gates, before policies, before governance committees, the single most impactful governance investment is making AI decisions visible.

In most Saudi organizations that have deployed AI incrementally across departments, the people responsible for AI-driven processes cannot see, in real time, what the AI is actually deciding. They receive outputs — a route assignment, a risk score, a recommended action — without visibility into the reasoning, the confidence level, the historical accuracy of comparable decisions, or the alternatives the system considered and rejected. Governance built on top of this invisibility is governance without a foundation.

Visibility changes the dynamic immediately. When managers can see AI decisions in a dashboard, in a decision log, in a real-time alert system, they catch anomalies that would otherwise be invisible until they become incidents. They develop an operational intuition for what normal looks like, which makes abnormal detectable early. The NCA guidance requiring monitoring infrastructure for AI systems is partly motivated by exactly this dynamic: an AI system that cannot be monitored cannot be governed, regardless of the quality of the documents surrounding it.

Organizations that sequence transparency first — before implementing approval gates — consistently find that many of the approval gates they anticipated needing become unnecessary once visibility is established. Anomalies that would have required approval to prevent are caught and corrected at the monitoring layer. The governance architecture becomes lighter and more effective at the same time.

The Operator Problem

AI governance documentation that exists only for compliance purposes — written for auditors and ignored by the people who operate AI systems — is not governance. It is record-keeping.

The test for whether governance is operational is simple: can the logistics manager, the loan officer, the clinical administrator, or the operations director use the governance artefacts in their daily work? If the answer is no — if the incident response playbook is written in language that assumes a compliance officer rather than an operations manager will be the one responding — the governance has not been integrated into operations. It exists as a parallel system that gets consulted after incidents, not before.

The organizations across the Kingdom that have made this transition treat front-line operators as the primary audience for governance artefacts. Incident response procedures are written for the person who will use them under pressure, not for the consultant who documented them. Monitoring dashboards surface the metrics that matter to the person running the operation, not the metrics that appear in the audit report. Policy documents answer the questions that operators actually ask: am I authorized to override this recommendation? What do I do if the system produces an output that does not look right? Who do I call? Audit compliance becomes a natural consequence of governance that functions in practice, not the primary objective of governance that exists on paper.

What Consistently Fails

Certain governance approaches produce results that are reliably poor, regardless of the quality of the underlying thinking.

Heavy approval bottlenecks — requiring senior sign-off on every AI deployment or every AI decision type — consistently eliminate the efficiency that AI is deployed to create, without producing proportionate safety improvements. Tiered approval processes, in which routine deployments follow a standard track while high-risk applications receive additional review, produce better safety outcomes and better deployment outcomes than uniform friction applied to everything.

Framework copy-paste produces governance that is structurally plausible and operationally inert. ISO 42001, SDAIA's AI Ethics Principles, and EU AI Act-adjacent frameworks are valuable inputs to governance design, not substitutes for it. Every organization's AI risk profile is shaped by its specific systems, deployment contexts, data assets, and regulatory obligations. Borrowed structure needs to be filled with substance that reflects those specifics; without that, the framework describes a governance function that no one inside the organization has ever actually operated.

Annual-review governance is perhaps the most widespread failure mode. AI systems change continuously — through model drift as real-world data distributions diverge from training data, through system updates that alter model behavior, through evolving deployment contexts. An annual review cycle creates a governance gap that grows steadily between reviews and closes with a retrospective exercise rather than a genuine ongoing assessment. Continuous monitoring with periodic deep reviews produces consistently better results.

Separating technical from organizational governance produces two parallel systems that rarely speak to each other. A monitoring dashboard without an escalation path is a data display. A governance policy without monitoring infrastructure is an aspiration. Effective governance integrates them so that technical systems surface the information governance processes need, and governance processes respond to what technical systems detect.

The Regulatory Outlook

For organizations operating in Saudi Arabia and the Gulf, the direction of regulatory travel is not ambiguous. AI governance is becoming a formal compliance requirement across an increasing number of sectors. SDAIA has signaled intent to move from guidance to certification and enforcement. SAMA's AI guidance already carries audit implications in the financial services context. The NCA's frameworks are extending progressively into new sectors.

The organizations that build governance capacity proactively are consistently better positioned when compliance windows arrive than those that build under deadline. The cost of emergency governance remediation — conducted under time pressure, without the organizational learning that comes from building governance functions gradually — is real and substantial. The institutional capability built through proactive investment has value that extends well beyond any single regulatory deadline.

The core principle that holds across every context is the same: governance that functions in practice, that operators use, that surfaces information, that produces responses when things go wrong, is the only governance that produces the outcomes it is designed to produce. Everything else, however well-documented, is the appearance of governance rather than the thing itself.

Published by PeopleSafetyLab — AI safety and governance research for KSA organizations.

N

Nora Al-Rashidi

AI governance researcher specialising in regulatory compliance for organisations in Saudi Arabia and the GCC. Examines how SDAIA, SAMA, and the NCA's overlapping frameworks interact — what that means for risk, audit, and board-level accountability.

Share this article: