The Numbers Behind the Reversal
The statistics have converged from enough independent sources to describe a genuine trend, not an outlier reaction. Flexera’s 2025 State of the Cloud Report found that 20% of workloads originally migrated to public cloud have already been repatriated to private or on-premises environments. IDC research from June 2024 found 80% of enterprises expected to repatriate some compute or storage workloads within 12 months. The Barclays CIO survey from Q4 2024 found 83% of enterprise leaders planned public cloud workload shifts. OpenText and Nutanix’s 2026 joint survey found 67% of organizations had already repatriated some workloads, with 87% planning to do so within 12–24 months.
What makes these statistics coherent rather than contradictory is the distinction they all maintain: repatriation does not mean abandoning cloud. Only 8% of organizations in IDC’s October 2024 survey planned to move entire workload portfolios off public cloud. What the 80–87% of organizations are doing is selective repatriation — moving specific workload categories back to private infrastructure while retaining public cloud for other categories. The outcome is not “back to on-premises” but “deliberately hybrid.”
The named corporate examples illustrate the economics. 37signals — the software company behind Basecamp and HEY — was spending $3.2 million annually on AWS. It projected $7 million in savings over five years by moving its primary workloads back to owned hardware. GEICO moved large portions of its compute off public cloud and achieved 50% per-core cost reduction. Broadcom achieved 40–50% lower total cost of ownership for steady-state workloads and saved over $10 million by moving database workloads back on-premises. These are not small-margin improvements — they are structural cost differences on production workloads that have reached steady-state operational maturity.
Three Drivers That Differentiate the 2026 Wave from Earlier Reversals
Cloud repatriation is not new. Enterprises have moved workloads back on-premises before, typically after failed migrations or during cost-cutting cycles. What distinguishes the 2026 wave is that three drivers are operating simultaneously and reinforcing each other in a way that creates structural, not cyclical, pressure.
Driver one: AI data gravity. The arrival of enterprise AI at scale has changed the economics of data movement. AI pipelines require enormous data volumes to flow between storage, compute, and inference endpoints. Egress fees — the charges cloud providers levy for moving data out of their networks — that were manageable for traditional application workloads become prohibitive when data moves are measured in petabytes per month. Deloitte’s analysis quantifies this: on-premises AI infrastructure delivers 50%+ cost savings over three years compared to cloud API pricing for steady-state AI workloads. Enterprises running large language model inference on proprietary data are discovering that the egress and API costs of cloud-based AI are structurally higher than the hardware amortization cost of running equivalent inference on-premises.
Driver two: sovereignty regulation. IDC data shows that 57% of IT leaders now require single-country infrastructure for at least some data categories. The EU’s GDPR has been supplemented by the Digital Operational Resilience Act (DORA), which took effect in January 2025 and imposes specific requirements on financial services firms regarding where data is processed and how ICT risk is managed. The UK’s Cyber Security and Resilience Bill, regulations in Saudi Arabia, Brazil, and India, and Algeria’s own Law 18-07 all impose constraints that hyperscaler “region” configurations satisfy imperfectly — because the operating entity remains a US-headquartered corporation subject to US jurisdiction regardless of where its servers are physically located. Enterprises that previously accepted “data stored in EU” as sufficient compliance are re-examining that conclusion under stricter regulatory scrutiny.
Driver three: modern hardware economics. The economics of owning versus renting compute have shifted. Server prices have declined. Modern server hardware has lower power consumption per compute unit than five years ago. The residual value of owned hardware — especially GPU servers — has improved. And crucially, the software management layer for on-premises infrastructure has dramatically improved: Kubernetes, Terraform, and hypervisor automation tools have reduced the operational burden of managing owned infrastructure to levels that smaller IT teams can sustain. The argument that cloud is operationally simpler than on-premises was true in 2015; it is less true in 2026.
Advertisement
What CIOs and Infrastructure Architects Should Do
1. Run a Workload-by-Workload Cost-to-Serve Analysis Before Your Next Cloud Contract Renewal
The single most valuable action a CIO can take in 2026 is a granular cost-to-serve analysis of each cloud-hosted workload before the annual renewal cycle. Most enterprises know their total cloud spend; few know the unit economics of individual workloads. The analysis should capture, per workload: compute cost (instances, hours), storage cost (GB-months, storage tier), egress cost (data transfer out), and associated labor cost (DevOps time attributable to maintaining the workload). Compare this to the capital cost of owning equivalent hardware, amortized over five years, plus power, space, and management labor. For workloads that have been running unchanged for 18+ months — what the industry calls “steady-state” — the on-premises economics are typically 30–60% cheaper. For variable, burst-capacity workloads, cloud typically remains the better option. The analysis tells you which workloads to repatriate and which to keep in cloud — not a blanket policy either way.
2. Repatriate Steady-State AI Inference Before Your Cloud AI Bill Compounds
The Deloitte 50%+ savings figure for on-premises versus cloud AI is not hypothetical — it reflects the specific economics of steady-state LLM inference. An enterprise running 10,000 queries per day against a proprietary fine-tuned model through a hyperscaler API pays per-token pricing that compounds rapidly. The same inference workload running on an owned or co-located GPU server pays for the server (amortized), electricity, and management — typically at 40–60% of the per-token API equivalent for any workload that runs consistently, rather than in bursts. The trigger for repatriation is predictability: when your AI query volume is predictable enough to plan hardware procurement around (within ±30%), on-premises inference economics are likely superior. When volumes are unpredictable or seasonal, cloud API remains the right choice.
3. Design Your Hybrid Architecture Around a Data Gravity Map, Not a Technology Preference
The most common hybrid architecture mistake is designing the split based on technology preferences (“we use VMware on-premises and Kubernetes in cloud”) rather than data gravity. Data gravity describes where data naturally accumulates — and workloads gravitate toward the data, not the other way around. If your AI training data lives in an on-premises data warehouse because regulatory constraints prevent it from leaving your jurisdiction, your AI training workload also belongs on-premises or in a sovereign cloud that meets the same compliance bar. Mapping data gravity first — identifying where each major dataset legally must reside — creates a defensible architecture for hybrid deployment that is compliance-correct by design rather than by accident. The DORA compliance requirement of documented exit strategies and verifiable encryption key control is only achievable if the architecture was designed with data location in mind from the start.
4. Quantify Your Exit Costs Before Signing Multi-Year Cloud Commitments
One factor driving the 2026 repatriation wave is that enterprises signed 3–5 year cloud commitment agreements in 2021–2023 and are now hitting renewal cycles with a clearer view of the economics. Before signing the next multi-year committed-use agreement, quantify the exit cost: what would it cost to migrate these workloads off this cloud at contract end? The cost has three components — data egress (the cloud provider’s charge for extracting your data), migration labor (engineering time to refactor workloads for a new destination), and productivity loss during cutover. For large enterprise workloads, these exit costs can exceed $1 million and take 6–12 months of engineering effort. This is the structural lock-in that makes cloud contracts stickier than they appear on annual cost reviews. Quantifying it explicitly enables more honest procurement decisions.
5. Treat Private Cloud and Co-Location as First-Class Options in Every Procurement Cycle
The procurement culture in most enterprise IT organizations has defaulted to “evaluate public cloud first” for the past decade. This default should be replaced with “evaluate all three options” — public cloud, private cloud (on-premises or in dedicated co-location), and hybrid — for every workload procurement decision. The tools to manage private cloud at enterprise scale (VMware, Red Hat OpenShift, Nutanix) have matured to the point that operational complexity is not a differentiating disadvantage for mid-to-large IT organizations. The economics analysis alone — the workload-by-workload cost-to-serve — should drive the placement decision, not the procurement default.
The Correction Scenario
Cloud repatriation can fail, and it has failed at companies that approached it as a cost-cutting exercise without understanding the operational requirements. The correction scenario looks like this: an enterprise repatriates a workload to cut costs, discovers that on-premises management requires staffing a 24/7 NOC team it does not have, and incurs enough operational incidents to negate the cost savings within 18 months. The remedy is not to avoid repatriation — it is to ensure that operational readiness is a prerequisite, not an afterthought. Only repatriate workloads for which your organization has (or will build) the on-premises operational capability to match the SLA your cloud provider was delivering. The cost savings of repatriation are real and documented; the operational risk of repatriation is also real and documented. Both must be factored.
The enterprises that get hybrid right in 2026 will build a competency — the ability to rationally place workloads across public cloud, private infrastructure, and sovereign cloud based on cost, compliance, and operational fit — that will be a durable competitive advantage through the 2030s. The enterprises that treat repatriation as a trend to follow, rather than a calculation to run, will spend the next five years oscillating between cloud and on-premises without capturing the economics of either.
Frequently Asked Questions
Q: Does cloud repatriation mean enterprises are giving up on cloud entirely?
No. IDC data shows only 8% of organizations plan to move entire portfolios off public cloud. The overwhelming majority are executing selective repatriation — moving specific workload categories (steady-state compute, regulated data, AI inference) to private or sovereign infrastructure, while retaining public cloud for burst capacity, development tools, and international SaaS services. The outcome is a more sophisticated hybrid architecture, not a return to pre-cloud operations.
Q: Which workloads benefit most from repatriation in 2026?
Steady-state workloads that have been running unchanged for 18+ months — particularly database workloads, internal business intelligence, and predictable AI inference pipelines — show the strongest economics for repatriation. Burst-capacity workloads, globally distributed applications, and development/test environments typically still favor public cloud. The key criterion is predictability: workloads with predictable, consistent resource consumption are better candidates for owned or co-located infrastructure.
Q: How does DORA (Digital Operational Resilience Act) affect cloud repatriation decisions?
DORA, effective January 2025 for EU financial services firms, requires verifiable control over ICT risk — including documented exit strategies from cloud providers, encryption key control that does not depend on the cloud vendor, and ICT concentration risk management. These requirements make it more difficult for financial firms to treat hyperscaler cloud as the only infrastructure option, since single-vendor dependence on a US-headquartered hyperscaler creates concentration risk that DORA regulations specifically target.
—
Sources & Further Reading
- What’s Driving Cloud Repatriation in 2026? — MRC Productivity
- Cloud Repatriation in 2026: Costs, Compliance, and Control — Volico
- Cloud Strategy for 2026: The Year of Repatriation, Resilience, and Regional Rebalancing — Digitalisation World
- Cloud Repatriation Guide 2026 — Tasrie IT
- Cloud Repatriation Trends — Servers.com
















