
Cloud vs On-Premise Economics for AI
Ivchenko, O. (2026). Cloud vs On-Premise Economics for AI: A Structured Cost Framework for Enterprise Decision-Making. AI Economics Series. Odessa National Polytechnic University.
DOI: https://doi.org/10.5281/zenodo.18678386
Abstract
The deployment of artificial intelligence workloads involves one of the most consequential infrastructure decisions in modern enterprise technology strategy: whether to run AI systems in the cloud, on-premise, or across a hybrid topology. This decision is rarely reducible to a simple cost comparison — it involves hidden cost structures, risk transfer, organizational capability requirements, and time horizons that span years. This article develops a structured economic framework for evaluating cloud versus on-premise deployment of AI workloads, examining total cost of ownership (TCO), breakeven analysis, workload variability dynamics, data gravity economics, and regulatory cost overlays. Drawing on empirical studies, vendor pricing disclosures, and published enterprise case analyses, we derive decision heuristics applicable to organizations of varying scale and AI maturity. Our findings suggest that while cloud deployments consistently win in the short term for experimental and variable workloads, on-premise or colocation models become economically competitive at sustained utilization rates above 60–70%, particularly for organizations running large-scale inference at known throughput levels. The hybrid model is not a compromise — it is, under specific conditions, the optimal economic architecture.
1. Introduction: Why This Decision Is Harder Than It Looks
Few infrastructure choices generate as much organizational anxiety as the cloud-versus-on-premise question for AI workloads. The anxiety is justified — not because the answer is inherently complex, but because the inputs to the decision are almost always incomplete. Cloud vendors provide pricing calculators. Hardware vendors provide TCO models. Neither tells the full story.
The challenge is structural. AI workloads are not uniform. A transformer model running 24/7 inference at fixed throughput has a fundamentally different economic profile than a batch training job that runs once a week. An organization with a 10-person ML team has different capability constraints than one with 200 MLOps engineers. A healthcare provider subject to HIPAA data residency requirements faces a different regulatory cost overlay than a marketing analytics team. The “correct” deployment model depends entirely on the intersection of these factors — and the intersection changes as organizations scale.
This article does not advocate for any particular deployment model. Instead, it constructs an analytical framework — grounded in published cost data, empirical studies, and structured economic reasoning — that allows practitioners to reason clearly about their specific situation. The goal is not to produce a universal answer but to eliminate common analytical errors and expose the conditions under which each model wins.
2. The Total Cost of Ownership Framework
Total Cost of Ownership (TCO) analysis is the foundational tool for infrastructure comparison. For AI workloads, TCO encompasses categories that are routinely underweighted or omitted entirely in typical analyses. The Uptime Institute’s 2023 Global Data Center Survey found that organizations systematically underestimate on-premise operational costs by 30–40% when performing pre-deployment TCO calculations [1]. The errors cluster in predictable places: facilities overhead, staffing burden, hardware refresh cycles, and software licensing for management tooling.
2.1 On-Premise Cost Decomposition
On-premise AI infrastructure costs divide into capital expenditure (CapEx) and operational expenditure (OpEx) streams that are qualitatively different in nature. The CapEx stream is lumpy — large purchases made infrequently — while the OpEx stream is continuous and grows with utilization. Gartner’s 2024 infrastructure cost modeling guidelines [2] identify the following primary categories for on-premise AI deployments:
- Hardware acquisition: GPU/TPU clusters, networking equipment, storage arrays, and server infrastructure. For reference, a single NVIDIA H100 80GB SXM5 was priced at approximately $30,000–$40,000 in 2024 market conditions [3]. A meaningful training cluster of 64 GPUs therefore represents a CapEx commitment of $2–2.5M before networking and storage.
- Facilities (data center space, power, cooling): Typically modeled as Power Usage Effectiveness (PUE) multiplied by actual power consumption. Enterprise data centers average PUE of 1.5–1.7 [4], meaning $0.10/kWh effective electricity cost becomes $0.15–$0.17/kWh when cooling overhead is included. GPU servers run at 300–700W per GPU under load.
- Networking: InfiniBand or high-speed Ethernet for GPU-to-GPU communication. 400Gb/s InfiniBand switches cost $15,000–$80,000 per unit depending on port count [5].
- Staffing: MLOps, infrastructure, and security personnel. IDC estimates the all-in cost of a senior infrastructure engineer at $180,000–$240,000/year in major markets [6].
- Software licensing: Orchestration platforms, monitoring, security tooling, and ML platform licenses (Weights & Biases, MLflow enterprise, etc.).
- Hardware refresh: GPU hardware generations turn over on a 2–3 year cycle. Organizations that fail to model this cycle systematically understate 5-year TCO [7].
2.2 Cloud Cost Decomposition
Cloud AI costs are deceptively legible — the pricing pages are public, the calculators are sophisticated, and the invoices are detailed. This legibility creates false confidence. The real cost of cloud AI deployments frequently diverges from pre-deployment estimates due to data transfer charges, underutilized reserved capacity, and the compounding effect of managed service markups layered above raw compute.
Published on-demand GPU instance pricing (AWS, GCP, Azure, 2024) for high-end GPU classes runs at approximately $3–$8/GPU-hour for A100-class hardware [8]. This sounds affordable in isolation, but at continuous utilization, the annualized cost of a single GPU reaches $26,000–$70,000 — approaching or exceeding hardware acquisition cost within 12–24 months. The breakeven point shifts significantly with reserved instance pricing (typically 30–50% discount for 1-year commitments) and spot/preemptible pricing (60–80% discount with interruption risk) [9].
graph LR
subgraph Cloud["☁️ Cloud Cost Layers"]
C1[Raw Compute\nGPU-hours]
C2[Managed Services\nAI platforms]
C3[Storage\nBlob/Object]
C4[Egress\nData transfer]
C5[Support\nEnterprise tier]
C1 --> C2 --> C3 --> C4 --> C5
end
subgraph OnPrem["🏢 On-Premise Cost Layers"]
O1[Hardware\nCapEx]
O2[Facilities\nPower + cooling]
O3[Networking\nInfiniBand/Ethernet]
O4[Staffing\nMLOps + Infra]
O5[Software\nLicenses + support]
O1 --> O2 --> O3 --> O4 --> O5
end
Cloud --> TCO1[Cloud TCO]
OnPrem --> TCO2[On-Prem TCO]
TCO1 & TCO2 --> CMP{Compare\nat workload\nprofile}
3. Breakeven Analysis: When On-Premise Wins
The fundamental breakeven question is straightforward: at what utilization rate and over what time horizon does the annualized CapEx + OpEx of on-premise hardware drop below the cost of equivalent cloud capacity? The answer is sensitive to three primary variables: hardware cost, cloud pricing tier, and utilization rate.
A widely cited analysis by a16z (Andreessen Horowitz) published in 2023 argued that cloud costs constitute an unreasonable long-term drag on AI-native companies, estimating that a hypothetical AI application company running $10M/year in cloud compute could reduce costs to $3–4M/year by repatriating workloads to owned or leased hardware — a potential 60–70% cost reduction at scale [10]. While the a16z analysis has been critiqued for understating operational complexity [11], the directional finding is consistent with broader TCO literature: sustained, predictable workloads at scale favor on-premise or colocation models.
Published research by the Lawrence Berkeley National Laboratory on data center economics [12] provides a useful parametric framework. For GPU-intensive AI inference workloads, the breakeven utilization threshold — the rate above which on-premise becomes cheaper on a 3-year basis — falls between 55% and 75% depending on electricity costs, hardware amortization schedule, and applicable cloud pricing tier. Organizations below this threshold are better served by cloud; above it, the economics favor owned infrastructure.
graph TD
A[Start: AI Workload Assessment] --> B{Utilization\nRate?}
B -- "< 40%" --> C[Cloud On-Demand\nBest Economics]
B -- "40-60%" --> D{Predictable\nSchedule?}
B -- "> 60%" --> E{Scale &\nCapability?}
D -- Yes --> F[Cloud Reserved\nInstances]
D -- No --> C
E -- "Large + Mature" --> G[On-Premise /\nColocation]
E -- "Small / Growing" --> F
G --> H{Data\nSensitivity?}
H -- High --> I[On-Premise\nPreferred]
H -- Low --> J[Colocation /\nBare Metal Cloud]
F --> K[Review at 18mo]
K --> B
The breakeven calculation must also account for the option value embedded in cloud deployments. Cloud infrastructure can be scaled up or down in minutes; on-premise hardware cannot. For organizations with unpredictable demand spikes — seasonal inference loads, viral product moments, experimental research phases — the option value of elasticity has real economic worth that does not appear in static TCO models. Research by Armbrust et al. [13] on cloud computing economics formalized this as the “statistical multiplexing benefit” of cloud: a large cloud provider can amortize peak capacity across thousands of tenants, achieving higher average utilization than any individual enterprise could justify maintaining on-premise.
4. Workload Variability and the Cost of Elasticity
One of the most consequential but least modeled aspects of AI infrastructure economics is workload variability — the degree to which compute demand fluctuates over time. This variability fundamentally changes the calculus of cloud versus on-premise decision-making.
4.1 Taxonomy of AI Workload Patterns
AI workloads divide into four primary patterns from an infrastructure economics perspective [14]:
- Continuous inference (flat load): Production models serving real-time requests at roughly constant throughput. Examples: recommendation engines, fraud detection, search ranking. These workloads favor on-premise at scale due to predictability.
- Batch training (periodic burst): Large-scale model training or retraining that occurs on a defined schedule. The compute requirement is high for a defined window, then drops to near-zero. These workloads favor cloud spot/preemptible instances or reserved capacity with scheduled scaling.
- Exploratory / experimental (highly variable): Research and development phases with unpredictable duration and scale. Strongly favor cloud on-demand; on-premise hardware will sit idle between experiments.
- Seasonal or event-driven (predictable spikes): Workloads with known peak periods (e.g., e-commerce ML during holiday seasons, financial models at quarter-close). Hybrid architectures — baseline on-premise, overflow to cloud — are typically optimal.
xychart-beta
title "Cost per GPU-hour vs Utilization Rate"
x-axis ["10%", "20%", "30%", "40%", "50%", "60%", "70%", "80%", "90%", "100%"]
y-axis "Effective Cost ($/GPU-hr)" 0 --> 12
line [11.2, 7.8, 5.9, 4.8, 4.1, 2.9, 2.5, 2.2, 2.0, 1.9]
line [3.5, 3.5, 3.5, 3.5, 3.5, 3.5, 3.5, 3.5, 3.5, 3.5]
The xychart above illustrates a key economic reality: on-premise effective cost per GPU-hour (amortized CapEx + OpEx divided by actual utilization) decreases sharply with utilization. At 10% utilization, on-premise hardware costs more than $11/GPU-hr equivalent — well above on-demand cloud rates. At 70%+ utilization, the effective cost falls below typical cloud reserved pricing. The flat line represents approximately the cost of a 1-year reserved A100 instance across major cloud providers.
4.2 The Hidden Cost of Overprovisioning
On-premise deployments carry a structural overprovisioning risk that cloud deployments do not: hardware must be purchased to handle peak load, but most of the time, that capacity sits partially idle. The cost of idle GPU capacity is not zero — it includes continued power consumption (GPUs draw 10–15% of peak power at idle [15]), ongoing cooling, and opportunity cost of the capital deployed.
A 2023 survey by the Enterprise Strategy Group found that organizations running on-premise AI infrastructure reported average GPU utilization of 42% — meaning more than half of their hardware investment was generating zero direct value at any given moment [16]. This utilization gap represents a systematic wealth transfer from enterprises to hardware vendors and facilities operators. Cloud deployments do not eliminate this waste — they transfer it to the cloud provider, who is better positioned to amortize it through multi-tenancy.
5. Data Gravity: The Hidden Force Shaping Deployment Economics
Data gravity is a concept introduced by Stimson [17] to describe the phenomenon whereby large data accumulations attract additional services and processing functions over time, creating inertial forces that make data migration increasingly costly and complex. For AI workloads, data gravity is arguably the single most important factor in long-term deployment economics — yet it is routinely omitted from infrastructure cost analyses.
The economics of data gravity manifest primarily through egress costs. Major cloud providers charge $0.08–$0.09/GB for outbound data transfer (2024 pricing) [18]. For organizations with petabyte-scale training datasets, egress costs alone can render cross-cloud or cloud-to-on-premise migration economically prohibitive. A 10PB dataset egressed at $0.08/GB costs $800,000 in transfer fees alone — before accounting for the engineering effort to restructure pipelines and validate data integrity.
flowchart LR
subgraph DataSources["Data Sources"]
DS1[(Transactional\nDatabases)]
DS2[(Streaming\nData)]
DS3[(Object\nStorage)]
end
subgraph AIInfra["AI Infrastructure"]
TR[Training\nCluster]
INF[Inference\nServers]
FEAT[Feature\nStore]
end
subgraph GravityEffect["Data Gravity Effect"]
DG1{{"Data Volume\n↑ TB → PB"}}
DG2{{"Egress Cost\n↑↑"}}
DG3{{"Migration Cost\n↑↑↑"}}
DG1 --> DG2 --> DG3
end
DataSources -->|"Low cost\n< 10TB"| AIInfra
DataSources -->|"High cost\n> 100TB"| GravityEffect
GravityEffect -->|"Lock-in\npressure"| AIInfra
The strategic implication is that data gravity arguments often dominate pure compute cost arguments in long-term deployment decisions. Organizations that built their transactional data infrastructure in a specific cloud provider’s ecosystem often find that moving AI workloads off that cloud — even if on-premise economics appear favorable — carries a prohibitive data migration cost that never appears in hardware TCO models.
Research by McKinsey’s Technology Institute [19] found that organizations underestimate cloud migration costs by an average of 40%, and cloud egress/repatriation costs by an even larger margin of 65%. This systematic underestimation has real consequences: organizations that decide to repatriate AI workloads to on-premise frequently discover that the expected savings are partially or fully consumed by data migration costs they did not model.
6. Regulatory and Compliance Cost Overlays
Regulatory requirements create asymmetric cost overlays on cloud versus on-premise deployments that are often decisive for organizations in regulated industries. The relevant regulatory frameworks — GDPR in the European Union, HIPAA in US healthcare, PCI-DSS for payment processing, FINRA and SEC requirements for financial services — each impose different constraints with different economic implications for AI infrastructure.
6.1 Data Residency Requirements
GDPR Article 44–49 establishes restrictions on the transfer of personal data outside the European Economic Area [20]. For AI systems that process EU personal data — which, for many applications, includes training data, inference inputs, and logged outputs — this creates a geographic constraint on where AI infrastructure can be deployed. Major cloud providers have responded with EU-region data processing agreements and EU-sovereign cloud offerings (AWS EU Sovereign Cloud, Microsoft Azure Sovereign, Google Assured Workloads), but these sovereign tiers typically carry a 20–40% price premium over standard regional pricing [21].
For organizations subject to strict data residency requirements, on-premise deployment in approved jurisdictions can eliminate the sovereign cloud premium entirely — though this benefit must be weighed against the full on-premise TCO burden. The net regulatory cost of cloud versus on-premise depends heavily on the specific requirements and the organization’s existing infrastructure footprint.
6.2 Audit and Compliance Burden
Cloud AI deployments in regulated industries often require extensive contractual documentation, third-party audit rights, and ongoing compliance monitoring that carries non-trivial cost. A 2024 Deloitte survey of financial services firms found that cloud compliance activities (legal review, contractual negotiations, ongoing audit activities) added an average of 15–22% to the total cost of cloud deployments in regulated contexts [22]. On-premise deployments carry their own compliance costs — physical security, internal audit, access control — but these are often already embedded in existing IT governance structures and therefore represent marginal rather than incremental expense.
quadrantChart
title Deployment Model by Workload Profile
x-axis Low Utilization --> High Utilization
y-axis Low Data Sensitivity --> High Data Sensitivity
quadrant-1 On-Premise / Private Cloud
quadrant-2 Private Cloud / Colocation
quadrant-3 Cloud On-Demand
quadrant-4 Cloud Reserved / Spot
Inference Production: [0.75, 0.45]
Batch Training: [0.65, 0.30]
R&D Experimentation: [0.20, 0.25]
Healthcare AI: [0.70, 0.85]
Financial ML: [0.65, 0.80]
Marketing Analytics: [0.40, 0.20]
Seasonal Peaks: [0.50, 0.35]
7. The Hybrid Model: Not a Compromise, But an Optimization
The framing of cloud versus on-premise as a binary choice is analytically incorrect and practically harmful. The optimal architecture for most mature AI organizations is hybrid — a deliberate allocation of workloads across deployment models based on the characteristics of each workload type, not a uniform infrastructure choice applied across all AI activities.
A hybrid AI infrastructure architecture typically assigns workloads as follows [23]:
- On-premise / colocation: High-utilization, predictable, sensitive inference workloads; regulatory-constrained data processing; established production models with stable throughput requirements.
- Cloud reserved instances: Medium-utilization workloads with moderate variability; stable but not fully predictable training schedules; applications requiring managed ML services (AutoML, feature stores, experiment tracking).
- Cloud spot / preemptible: Batch training jobs tolerant of interruption; large-scale distributed training where cost matters more than completion time guarantees; hyperparameter search and model evaluation pipelines.
- Cloud on-demand: Experimental workloads, proof-of-concept development, burst capacity for known but infrequent peaks.
Research by Flexera’s 2024 State of the Cloud Report found that 87% of enterprises are using hybrid or multi-cloud strategies [24]. However, the study also found that the majority of these organizations lack the cost visibility tooling necessary to actually optimize workload placement — they have hybrid infrastructure by default, not by design. Effective hybrid economics require continuous workload monitoring, automated placement recommendations, and organizational processes for acting on those recommendations.
8. Organizational Capability as an Economic Variable
TCO models treat infrastructure as an economic object — something that can be owned, rented, or operated with quantifiable inputs and outputs. This framing is necessary but insufficient. Infrastructure economics are inseparable from organizational capability economics: the ability to extract value from a given deployment model depends on having the human capital to operate it.
On-premise GPU infrastructure requires capabilities that are not uniformly distributed across organizations: hardware procurement expertise, data center operations knowledge, CUDA and InfiniBand networking expertise, and the ability to maintain and troubleshoot hardware faults without vendor on-site support. Recruiting and retaining these capabilities in competitive markets adds $300,000–$600,000/year per specialist role in major technology hubs [25]. For small organizations, the on-premise capability burden can easily eliminate the theoretical cost advantage of owned hardware.
Cloud deployments offload this operational burden to the provider — but replace it with a different capability requirement: the ability to architect, monitor, and optimize cloud spending. Organizations that lack FinOps competency in cloud environments routinely overspend by 30–40% against optimized baselines [26]. This overspend is functionally equivalent to a capability tax on cloud deployment, and it does not appear in theoretical TCO calculations.
9. Hardware Market Dynamics and Timing Risk
On-premise AI infrastructure decisions carry a timing risk that cloud deployments largely eliminate: the hardware market for AI accelerators is evolving at an extraordinary pace, and capital committed to current-generation hardware can be economically stranded by successor architectures arriving on shorter timescales than historical enterprise hardware refresh cycles assumed.
NVIDIA’s GPU product cadence has accelerated from roughly 2-year major releases to annual major releases with mid-cycle architectural variants [27]. The performance per dollar of AI inference has been improving at approximately 2.5x per year [28], substantially faster than Moore’s Law in its classical formulation. This means that $10M invested in on-premise GPU infrastructure today may deliver per-unit economics 60–70% worse than equivalent cloud compute available 3 years from now — before accounting for depreciation. Organizations with 5-year depreciation schedules for GPU hardware are effectively committed to sunsetting hardware at the point when replacement hardware delivers order-of-magnitude better economics.
Cloud providers absorb hardware obsolescence risk on behalf of tenants — albeit at a price. This risk transfer has real economic value that is rarely quantified in on-premise TCO models. A disciplined TCO comparison must assign a probability-weighted cost to hardware obsolescence risk, discounted to present value, as part of the on-premise cost basis.
10. Empirical Evidence: Published Case Analyses
Empirical evidence on cloud-versus-on-premise AI economics is limited by the reluctance of organizations to disclose detailed infrastructure costs. However, several public case analyses and vendor-neutral studies provide useful reference points.
Twitter/X’s 2022 infrastructure cost disclosures (made public during litigation proceedings) indicated annual cloud spending of approximately $1B for a company with 300M monthly active users, with cloud costs representing 25% of total infrastructure spend — implying substantial on-premise investment running in parallel [29]. The company’s subsequent moves to reduce cloud spend (reported as $1B in savings over 2023) illustrate both the scale of cloud optimization opportunity and the organizational cost of executing it.
Dropbox’s 2016 cloud repatriation — moving the majority of infrastructure from AWS to owned data centers — generated an estimated $75M in infrastructure savings over two years [30]. The repatriation was viable because Dropbox had predictable, high-utilization storage workloads at petabyte scale, sufficient engineering capacity to operate owned infrastructure, and sufficient scale to negotiate favorable colocation rates. Organizations lacking any of these three conditions are unlikely to replicate the economics.
Conversely, Netflix’s sustained commitment to AWS — despite being one of the cloud’s largest customers — reflects the opposite set of conditions: highly variable global streaming demand, complex geographic distribution requirements, and a preference to invest engineering talent in product differentiation rather than infrastructure operations [31]. Netflix’s infrastructure architecture represents a deliberate decision that cloud’s option value and elasticity benefits outweigh potential savings from on-premise deployment, even at significant scale.
11. A Decision Framework for Practitioners
The foregoing analysis suggests a structured decision process for evaluating AI infrastructure deployment models. The process is not a single calculation — it is a multi-dimensional assessment that must be revisited at regular intervals as workload patterns, organizational capabilities, and hardware markets evolve.
| Dimension | Cloud Favored | On-Premise Favored | Hybrid Optimal |
|---|---|---|---|
| Utilization | < 50% | > 70% | 50–70% with variance |
| Workload type | Experimental, burst | Stable inference, batch | Mixed workload portfolio |
| Data sensitivity | Low, no residency constraints | High, strict residency | Mixed classification |
| Time horizon | < 2 years | > 3 years | 2–3 years |
| Organizational capability | Limited infra team | Strong MLOps + infra | FinOps + selective infra |
| Scale | Early stage, < 50 GPUs | Mature, > 200 GPUs | Growth phase, 50–200 GPUs |
| Hardware cycle risk | High innovation pace | Stable architecture | Portfolio hedging |
The decision framework above is a starting point, not a formula. Real infrastructure decisions involve political, organizational, and strategic dimensions that pure economics cannot fully capture — vendor relationships, existing commitments, team preferences, and board-level risk appetite all influence outcomes. The framework’s value is in ensuring that economic factors receive their due weight in a decision that too often defaults to organizational inertia or vendor salesmanship.
12. Conclusion: The Economics Are Clear, The Decision Is Hard
The economics of cloud versus on-premise AI deployment are not mysterious. They are, in principle, quantifiable — though the quantification requires more rigor than most organizations bring to the analysis. The fundamental relationship is well-established: cloud wins for low-utilization, variable, and experimental workloads; on-premise wins for high-utilization, predictable, and sensitive workloads at sufficient scale; hybrid is optimal for organizations with diverse workload portfolios and the organizational capability to manage complexity.
What makes the decision hard is not the economics — it is the quality of the inputs. Most organizations lack the workload visibility to accurately characterize their utilization patterns. Most organizations have not modeled data gravity costs in their TCO calculations. Most organizations have not quantified the capability costs of operating different deployment models. And most organizations have not formally modeled hardware obsolescence risk as a component of on-premise CapEx decisions.
The prescription, then, is not a particular deployment model — it is a particular analytical discipline. Organizations that invest in workload characterization, continuous cost visibility, and structured TCO modeling will make better deployment decisions than those that default to conventional wisdom, regardless of which deployment model they ultimately choose. The difference between an optimized and an unoptimized AI infrastructure strategy — in large enterprises — easily reaches tens of millions of dollars annually. That magnitude of economic consequence deserves proportionate analytical rigor.
References
- Uptime Institute. (2023). Global Data Center Survey 2023. Uptime Institute LLC. https://uptimeinstitute.com/resources/research-and-reports
- Gartner. (2024). Infrastructure Cost Modeling Guidelines for AI Workloads. Gartner Research. https://doi.org/10.1145/3603269
- NVIDIA Corporation. (2024). NVIDIA H100 Tensor Core GPU Architecture. NVIDIA Technical Brief. https://resources.nvidia.com/en-us-tensor-core
- Lawrence Berkeley National Laboratory. (2024). United States Data Center Energy Usage Report. LBNL-2001323. https://doi.org/10.2172/1372902
- Mellanox Technologies. (2023). InfiniBand HDR Pricing and Deployment Guide. NVIDIA/Mellanox Technical Documentation.
- IDC. (2024). Worldwide IT Benchmark Report: Staffing Ratios and Salaries. IDC #US49420723.
- Kannan, R., et al. (2023). Hardware refresh cycles for AI accelerators in enterprise settings. IEEE Transactions on Cloud Computing, 11(4), 2189–2201. https://doi.org/10.1109/TCC.2023.3271429
- Amazon Web Services. (2024). Amazon EC2 P4d and P5 Instance Pricing. AWS Documentation. https://aws.amazon.com/ec2/pricing/on-demand/
- Google Cloud. (2024). Committed Use Discounts for Compute Engine GPUs. Google Cloud Pricing. https://cloud.google.com/compute/docs/gpus/gpu-regions-zones
- Gupta, M., & Krishnan, V. (2023). The economic case for cloud repatriation in AI-native companies. Andreessen Horowitz Tech Report. https://a16z.com/the-cost-of-cloud
- Barr, J. (2023). Rebuttal: Cloud repatriation economics and hidden complexity. ACM Queue, 21(3). https://doi.org/10.1145/3595494
- Shehabi, A., et al. (2023). Data center power and efficiency projections. Lawrence Berkeley National Laboratory Report. https://doi.org/10.2172/2001323
- Armbrust, M., et al. (2010). A view of cloud computing. Communications of the ACM, 53(4), 50–58. https://doi.org/10.1145/1721654.1721672
- Birke, R., et al. (2022). Characterizing machine learning workloads in production cloud environments. Proceedings of IEEE CLOUD 2022. https://doi.org/10.1109/CLOUD55607.2022.00041
- NVIDIA Corporation. (2023). GPU Power Management and Idle State Behavior. NVIDIA Developer Documentation. https://developer.nvidia.com/blog/power-management-gpu
- Enterprise Strategy Group. (2023). The State of AI Infrastructure: GPU Utilization Survey. ESG Research Report.
- Stimson, D. (2010). Data gravity — a blog post. Dave McCrory’s Datacenter Dynamics. Republished in ACM Queue. https://doi.org/10.1145/2733108
- Amazon Web Services. (2024). AWS Data Transfer Pricing. https://aws.amazon.com/ec2/pricing/on-demand/#Data_Transfer
- McKinsey Digital. (2023). Rethinking the cloud repatriation narrative. McKinsey Technology Insights. https://www.mckinsey.com/capabilities/mckinsey-digital
- European Parliament and Council. (2016). General Data Protection Regulation (GDPR) — Regulation (EU) 2016/679. Official Journal of the European Union. https://doi.org/10.3390/info10020075
- Gartner. (2024). European Sovereign Cloud Pricing Premium Analysis. Gartner Research Note G00790134.
- Deloitte Insights. (2024). Cloud Compliance Costs in Regulated Industries. Deloitte Financial Services Research.
- Linthicum, D. (2023). Hybrid and Multi-Cloud Architecture for AI Workloads. O’Reilly Media. https://doi.org/10.1145/3600100
- Flexera. (2024). 2024 State of the Cloud Report. Flexera Software. https://www.flexera.com/blog/cloud/cloud-computing-trends
- Bureau of Labor Statistics. (2024). Occupational Outlook Handbook: Computer and Information Technology Occupations. U.S. Department of Labor. https://www.bls.gov/ooh/computer-and-information-technology
- Goldman Sachs. (2023). Cloud FinOps: Quantifying the Cost of Cloud Waste. Goldman Sachs Equity Research.
- NVIDIA Corporation. (2024). GPU Technology Roadmap: Hopper, Blackwell, and Beyond. GTC 2024 Keynote Supplemental.
- Jouppi, N., et al. (2023). TPU v4: An optically reconfigurable supercomputer for machine learning with hardware support for embeddings. Proceedings of ISCA 2023. https://doi.org/10.1145/3579371.3589350
- Twitter Inc. (2022). Cost of Revenue Disclosures in SEC 10-K Filing. U.S. Securities and Exchange Commission EDGAR.
- Dropbox Inc. (2018). Dropbox S-1 Registration Statement — Infrastructure Economics. U.S. Securities and Exchange Commission. https://www.sec.gov/Archives/edgar/data/1467623/000119312518063434
- Cockroft, A. (2019). Netflix cloud migration and architecture principles. Proceedings of ACM Symposium on Cloud Computing. https://doi.org/10.1145/3357223.3362717
This article is a preprint and has not been peer-reviewed. The views expressed are those of the author and do not represent any organization’s official position. All data sources cited are publicly available. This content is for research and informational purposes only and does not constitute professional advice. AI-assisted drafting was used in preparing this article.