March 2026·7 min read

Why is my AWS / Azure bill so high — and what to do about it

Cloud costs spiral for predictable reasons. Here is what is actually driving your bill, and the levers you can pull to bring it back under control — without slowing your team down.

You moved to the cloud to save money and move faster. Now you're staring at an invoice that's climbed every quarter, and nobody in the business can give you a straight answer about why.

You're not alone. Cloud cost overruns are one of the most common problems we see — and they're almost always fixable. But first you need to understand what's actually driving the bill.

The scale of the problem

Flexera's annual State of the Cloud report consistently finds that organisations estimate they waste around 28–32% of their cloud spend. In practice, for teams without active cost governance, the figure is often higher. The FinOps Foundation puts average cloud waste at closer to 30% of total spend — a number that has held stubbornly consistent for several years despite cloud providers building increasingly sophisticated cost management tooling.

The implication: for most organisations, a meaningful portion of the cloud bill is recoverable without any impact on capability. The challenge is knowing where to look.

The four reasons cloud bills spiral

1. Nobody owns it

On-premise infrastructure had a budget owner — usually IT. Cloud spend is different. It's spread across teams, charged by the minute, and easy to provision without anyone senior signing off. A developer spins up a test environment and forgets to tear it down. A data pipeline runs nightly against a dataset three times bigger than it was six months ago. A third-party integration starts making twice as many API calls after an update.

None of these are dramatic decisions. They're the accumulation of small choices nobody was incentivised to question.

The FinOps Foundation's 2024 State of FinOps report found that lack of cost ownership and accountability remains the top challenge for cloud financial management teams — ahead of tooling, forecasting accuracy, and engineering capacity. The problem isn't primarily technical. It's organisational.

2. You're paying for capacity you don't use

Cloud providers make it easy to overprovision. It feels safer to run a larger instance than to risk a performance issue. Reserved instance commitments get made and then the workload changes. Auto-scaling is configured with overly generous upper limits that regularly get hit for no good reason.

Right-sizing — matching your actual compute and storage to what workloads genuinely need — is one of the highest-return activities in cloud management. Flexera's data consistently shows that 20–40% of compute spend can be eliminated through right-sizing without any performance impact. For storage, the figures are often higher, particularly where teams are retaining data in hot storage tiers long past the point of active use.

3. Data transfer costs are invisible until they're not

Moving data between regions, out of the cloud to the internet, or between services within the same provider all carries a cost that's easy to miss at small scale and significant at large scale. As data volumes grow and architectures get more complex, transfer costs become a material line item that nobody anticipated when the architecture was designed.

This is especially true for organisations running hybrid architectures — part on-premise, part cloud — or using multiple clouds for different workloads. Every data movement between environments is a cost that accumulates quietly in the bill.

4. AI services are a new category of cost that most teams aren't tracking

If your teams are using managed AI services — foundation models, vector databases, embedding APIs, managed ML training — you're in new territory. These services don't follow the same pricing logic as compute and storage. Token-based pricing, per-request charges, and model training costs can scale non-linearly in ways that surprise even experienced engineers.

Gartner's analysis projects that AI-related workloads will account for over 60% of cloud infrastructure spend within a decade. Most organisations have no governance framework for this yet. The FinOps Foundation's 2024 report identified AI/ML cost management as the fastest-growing area of concern among practitioners — and the area where existing tooling and practices are least mature.

What to do about it

Start with visibility. You can't optimise what you can't see. Cloud providers offer native cost management tools — AWS Cost Explorer, Azure Cost Management, GCP Billing — but they require configuration and someone who knows what they're looking at. Even the native tooling will surface the big wins if you use it properly. A cost anomaly detection policy costs nothing to configure and can catch unexpected spend within hours rather than at month-end.

Tag everything. Cost allocation tags let you attribute spend to teams, projects, environments, and workloads. Without them, your bill is a lump sum. With them, it's a conversation you can have with the people generating the spend. The FinOps Foundation consistently identifies tagging coverage as the single most impactful first step — not because it reduces costs directly, but because it creates the visibility that makes everything else possible.

Separate production from everything else. Development, testing, and staging environments don't need to run at production scale 24 hours a day. Automated shutdown schedules for non-production environments are one of the easiest quick wins available — typically saving 60–70% of non-production compute costs with zero impact on developer productivity.

Review your commitments. Reserved instances and committed use discounts can deliver 30–60% savings compared to on-demand pricing — but only when the underlying workloads are stable and well understood. Review your commitment portfolio at least quarterly. Many organisations are paying for reserved capacity that no longer maps to how their workloads actually run.

Build a FinOps practice, not a one-off review. Cost optimisation isn't a project. It's an ongoing discipline. The FinOps Foundation defines the practice as "a cultural shift that enables teams to take ownership of their cloud usage." The organisations that control cloud spend well have someone — or a small team — whose job it is to watch the numbers, investigate anomalies, and work with engineering teams to build cost-awareness into how they work.

The shift that changes the economics

Working with a major travel and hospitality operator, we found that cloud costs had grown by over 200% in 18 months — not because the business had grown proportionally, but because nobody had governed the spend as the architecture evolved. The data and AI workloads had scaled; the oversight hadn't.

What struck us was that the FinOps capability needed to govern this — cost attribution, anomaly detection, commitment optimisation, AI spend forecasting — was the kind of thing that large enterprises used to build entire teams around. It was genuinely inaccessible to most organisations without significant headcount.

That's changed. The tooling, the cloud provider programmes, and the agentic AI that can now monitor and flag spend anomalies in real time means a well-governed cloud environment is no longer a large-enterprise luxury. A lean team with the right setup can maintain better cost visibility than a 50-person IT department was managing five years ago.

The organisations that get this right aren't just cutting costs. They're redirecting that spend toward the AI capabilities that actually generate competitive advantage — and with cloud co-funding programmes available for most AI workloads, the net cost of building intelligence is lower than most finance teams realise.

A word on reseller relationships

If you're buying cloud direct, you're paying list price and navigating funding programmes without support. A good cloud reseller should be doing more than managing your billing — they should be proactively flagging inefficiencies, helping you access co-funding and credit programmes, and bringing commercial levers that direct customers don't have.

If your current reseller's only communication is a monthly invoice, it's worth asking whether the relationship is delivering what it should.


If your cloud bill has been heading in the wrong direction and you're not sure why, the first step is usually a structured review of where the spend is actually going. We do this as part of our Cloud FinOps & AI Cost Management work — and in most cases, the savings we identify more than cover the cost of the engagement.

Talk to us about your cloud spend →


Sources

About the authors

DH

Daren Howell

Founder, CrewCreateAI

20+ years delivering AI and data programmes for global publishers, financial services firms, travel operators, and consumer brands. I've sat on both sides of the cloud invoice — as the person approving it and the person tasked with explaining why it keeps going up.

CM

CrewMate

AI Research Agent, CrewCreateAI

CrewMate draws on published research, technology documentation, industry analysis, and publicly available case studies to help identify patterns and strengthen every post.

Want to talk through what this means for your business?

Our AI Opportunity Scan is at no cost. In a few hours we'll give you a clear picture of where to focus — and what to do first.

Book Your Free Scan