Your AWS Bill Is Lying to You — Here's How LLMs Can Help You Read It
- 8 minutes read - 1647 wordsIt usually starts with a Slack message from finance.
“Hey, can someone explain why AWS spend jumped 40% this month?”
And then begins the archeological dig: Cost Explorer tabs, spreadsheets, Slack threads with the platform team, someone pulling CloudWatch metrics for a service that may or may not still exist. Three hours later, you have a partial answer and a lingering feeling that you’re only seeing the tip of the iceberg.
Sound familiar? You’re not alone. AWS cost management is genuinely hard — not because the tools don’t exist, but because the signal is buried in noise and the people who understand the bill aren’t always the people who have the context to fix it.
That’s exactly the problem Large Language Models (LLMs) are uniquely suited to solve.

Why AWS Costs Are So Hard to Control
AWS has over 200 services, operates across 30+ regions, and bills you on dozens of dimensions — compute hours, API calls, data transfer, storage tiers, license fees. A mid-sized company running a handful of production workloads can easily have 50,000+ line items in their monthly Cost and Usage Report.
AWS gives you tools to navigate this — Cost Explorer, Budgets, Compute Optimizer, Trusted Advisor — and they’re genuinely useful. But they share a common limitation: they surface data, not decisions.
The three failure modes we see again and again:
1. You’re always looking backward. By the time Cost Explorer shows you a spike, the damage is done. Budget alerts fire after the threshold is crossed. You’re managing a post-mortem, not preventing one.
2. The insights live in silos. Your CloudWatch metrics are in one place. Your billing data is in another. Your deployment history is in CI/CD. Connecting these dots to understand why a cost changed requires manual correlation across tools and teams.
3. The people who can act don’t know the data, and the people who know the data can’t act. Finance understands the budget. Engineers understand the architecture. Neither can fully read the other’s screen. The result is slow, frustrating handoffs — and costs that stay higher than they need to be.
Enter the LLM: Turning Billing Data into Plain English
Here’s what changes when you put an LLM between your AWS billing data and your team.
Ask Questions Like a Human
Forget filter dropdowns and date range pickers. Forget writing Athena SQL against your CUR bucket. You ask a question the way you’d ask a colleague:
“Which services drove the biggest cost increase last month?”
“Are there any EC2 instances in the dev account that haven’t had significant traffic in the past two weeks?”
“What would our bill look like if we moved all dev workloads to Spot instances?”
The LLM understands the intent, pulls the right data, and returns a clear answer in seconds. No query language. No BI tool training. No ticket to the data team.
Get a Root Cause, Not Just an Alert
AWS Cost Anomaly Detection will tell you there was a spike. An LLM will tell you why:
“The 340% increase in data transfer costs on May 3rd correlates with a deployment in us-east-1. Cross-region replication to eu-west-1 was not previously enabled for this service — this appears to be a configuration change introduced in the 2:14 PM deployment.”
That single sentence saves hours of investigation and points directly at the fix.
Rightsizing That Actually Gets Adopted
The dirty secret of cloud rightsizing is that most recommendations never get implemented. Engineers don’t trust them, don’t have time to validate them, or don’t understand the business context behind them.
An LLM-powered recommendation reads differently:
“14 instances in the dev account are m5.2xlarge but average 4% CPU and 12% memory over 60 days. Downgrading to m5.large would save approximately $2,800/month with no expected performance impact. These are non-production workloads with no SLA commitments.”
Context. Specificity. A clear action. That’s a recommendation an engineer will actually act on.
Monthly Reports That Tell a Story
Instead of a dashboard nobody reads, the same data becomes a narrative your entire company can understand:
“Your April AWS spend was $142,000 — up 18% from March. The primary driver is Amazon Bedrock usage, which tripled following your AI feature launch. EC2, RDS, and S3 costs are flat. There are three Reserved Instance purchases that could reduce your baseline by ~$21,000/month, outlined below.”
Finance gets a narrative. Engineering gets priorities. Everyone’s working from the same page.
Meet Optix-ai: Open-Source LLM-Powered Cost Intelligence
Optix-ai is an open-source project by Techtious built specifically for this problem. It connects directly to your AWS environment, ingests your real billing and utilization data, and pairs it with an LLM backend to deliver the kind of intelligence described above — all running in your own infrastructure.
No SaaS subscription. No billing data sent to third-party servers. No per-seat pricing.
What It Does
- Pulls live data from Cost Explorer, CloudWatch, and your Cost and Usage Report (CUR) via Athena
- Identifies idle resources, oversized instances, unused reservations, and data transfer inefficiencies
- Surfaces rightsizing recommendations for EC2, RDS, ECS, and Lambda, ranked by estimated monthly savings
- Provides a conversational chat interface so anyone on your team can query costs in plain English
- Works with OpenAI, Anthropic Claude, or locally-hosted models via Ollama — you choose the LLM backend
- Exports saving plans as Jira tickets or Slack messages to close the loop between insight and action
Why Open Source Is the Right Call for This
Cloud billing data is sensitive. It reveals product strategy, scaling velocity, and where your engineering investment is concentrated. You should not be shipping that to a third-party SaaS platform you don’t fully control.
With Optix-ai, the data stays in your AWS account. The LLM calls go to whichever endpoint you configure — including fully local models if data sovereignty is a requirement. And because it’s open source, you can extend it: add support for new AWS services, plug in internal cost allocation logic, or build custom dashboards on top of the API.
Getting Started in Minutes
# Clone the repo
git clone https://github.com/techtious/optix-ai
# Set up your environment
cp .env.example .env
# Add your AWS_PROFILE, LLM_PROVIDER, and API keys
# Launch with Docker Compose
docker compose up
Once it’s running, you get a web UI with a cost breakdown view, a chat interface, a recommendations panel, and export options. First meaningful insight typically takes under five minutes from a cold start.
The Quick-Win Checklist: What LLMs Find First
If you’re starting from scratch, here are the highest-ROI opportunities an LLM-assisted tool like Optix-ai will surface almost immediately:
| # | What to Look For | Why It Matters | Typical Saving |
|---|---|---|---|
| 1 | EC2 instances with < 15% avg CPU over 30 days | Pure waste — pay for what you use | 20–40% on compute |
| 2 | Unattached EBS volumes and unused Elastic IPs | Forgotten resources that keep billing | $5–50/resource/month |
| 3 | S3 objects untouched for 90+ days in Standard tier | Cold data in the wrong storage class | Up to 80% on storage |
| 4 | NAT Gateway data transfer across regions | Often a misconfiguration, not intentional | 30–60% on transfer |
| 5 | Stable workloads with no Savings Plan coverage | Leaving committed-use discounts on the table | 20–30% on covered spend |
| 6 | Resources with no cost allocation tags | Can’t optimize what you can’t attribute | Drives accountability |
| 7 | No anomaly alerts configured | Next spike will be a surprise | Prevents bill shock |
An LLM doesn’t replace this checklist — it automates the discovery step, explains every finding in context, and tells you which ones to prioritize. What would take a FinOps engineer a full day of investigation takes Optix-ai about three minutes.
How Optix-ai Is Wired Together
┌──────────────────────────────────────────┐
│ Your AWS Accounts │
│ Cost Explorer API │ CloudWatch Metrics │
│ CUR (S3 + Athena) │ EC2 / RDS Metadata │
└──────────────┬───────────────────────────┘
│
▼
┌──────────────────────────────┐
│ Optix-ai Core │
│ Data Ingestion & Caching │
│ Recommendation Engine │
│ Conversational Query Layer │
└──────────────┬───────────────┘
│
▼
┌─────────────────────────────────┐
│ LLM Backend │
│ OpenAI GPT-4 │ Claude │ │
│ Ollama (local, air-gapped) │
└──────────────┬──────────────────┘
│
▼
┌──────────────────────────────┐
│ Web UI / API / CLI │
│ Engineers + Finance │
│ Jira / Slack Exports │
└──────────────────────────────┘
The design keeps your data layer and your intelligence layer separate. Swap out the LLM backend without touching the ingestion pipeline. Add new AWS data sources without changing the chat interface.
What Teams Typically See in the First 90 Days
The numbers aren’t theoretical — they reflect what happens when you move from manual cost reviews to continuous, AI-assisted visibility:
- 15–35% reduction in monthly AWS spend, driven mainly by rightsizing and idle resource cleanup
- 70% less time spent on monthly cost reviews — from multi-hour analyst sessions to a 15-minute scan
- Faster engineer buy-in — recommendations in plain English tied to specific resources get acted on; abstract dashboards don’t
- Finance and engineering finally speaking the same language — both teams querying the same tool, getting consistent answers
The biggest unlock isn’t the savings themselves — it’s the speed of the feedback loop. When you can ask a question and get a useful answer in 30 seconds, you ask more questions. You catch things earlier. You stop waiting for the monthly bill to tell you what happened.
Start Small, See Fast Results
You don’t need to overhaul your entire FinOps practice to benefit from LLM-assisted cost optimization. A single starting point:
- Deploy Optix-ai in one AWS account — your staging or dev environment is a great low-risk start
- Let it run for a week to build up context on your usage patterns
- Ask it one question: “What are the top three things I could do right now to reduce this account’s bill?”
That first answer usually pays back the setup time many times over.
Optix-ai on GitHub: github.com/techtious/optix-ai Learn more: techtious.com
Cloud costs aren’t going down on their own. But with the right intelligence layer, they don’t have to be a mystery either.