Protect online privacy from the very first click
Your digital footprint begins long before you understand what it means. “Free” Big Tech inboxes like Gmail scan your emails to fuel advertising, personalize content, and build data profiles. Proton Mail offers truly “free” email. Free from data profiling. Free from tracking. Free from ads. And free to use.
AI
Shared AI Infrastructure Is Where Chargeback Models Break Down
A company with three engineering teams sharing an AWS environment ran showback-only for a year, got tagging coverage to 87 percent, then moved to a hybrid chargeback model. Within one quarter, the Data Science team had implemented spot instance usage for non-critical training jobs and cut their GPU spend by 31 percent. Product Engineering started shutting down staging environments on weekends. Total cloud waste dropped without a mandate because the costs were now visible and attributed to each team's actual budget. (Holori)
That outcome is the intended arc of showback and chargeback working as designed. The model is straightforward for infrastructure that maps cleanly to a single owner.
It becomes substantially more complicated when the infrastructure in question is a shared AI inference endpoint serving three different products, or a centrally managed embedding pipeline that multiple business units depend on.
The Attribution Problem Unique to AI
Showback and chargeback both depend on being able to answer the question: who consumed this? For a dedicated compute instance running one team's workload, the answer is straightforward. For shared AI infrastructure, it is not.
When a central ML platform team operates an inference endpoint that multiple products consume, cost allocation requires knowing how much each product consumed, not just the aggregate cost of the endpoint. That requires request-level telemetry, rather than just resource-level tags. Traditional chargeback models that rely on infrastructure tagging cannot produce that attribution without instrumentation at the application layer.
CloudZero's analysis of AI cost attribution describes this as a structural problem: when multiple teams route requests through a single model endpoint, accurate allocation requires request-level telemetry, not just resource-level tags. Their CostFormation approach allocates shared costs including untagged resources by using usage signals rather than relying on complete tag coverage. (CloudZero) The implication is that AI chargeback requires more sophisticated attribution logic than traditional cloud cost allocation, and building that logic before rolling out chargeback is not optional, it is what makes the chargeback numbers defensible.
Agentic workflows introduce a related complexity. A single business process executed by an AI agent may generate cost events across an orchestration service, a compute instance, external API calls, storage operations, and a retrieval database. Those events appear in several billing line items across different service categories. Aggregating them into a coherent per-workflow cost that can be attributed to a team or product requires workflow-level instrumentation that goes beyond what standard FinOps tooling handles today.
The Sequencing That Makes Chargeback Work
The evidence for starting with showback before chargeback is consistent across implementations. Teams that did not understand their spending patterns before being charged for them tend to respond with confusion and contest the allocations, rather than optimizing their usage. The showback period is not wasted time, it surfaces tagging gaps, catches misattributed costs, and builds the organizational trust that makes chargeback numbers credible when they land in department budgets.

A reasonable benchmark is 6 to 18 months in showback mode before transitioning to chargeback for well-understood infrastructure. For AI workloads, the showback period also serves as the window for building the attribution instrumentation that AI-specific chargeback requires. A team that has 12 months of inference telemetry mapped to product surfaces, with allocation logic validated against known usage patterns, is in a materially better position to launch chargeback than one that has only billing data and tags.
For shared infrastructure, the hybrid model is the most commonly adopted approach in mid-to-large enterprises because it balances precision with practical simplicity. (Holori) Teams understand what they can directly influence through their architectural and operational choices, and the platform fee is treated as a utility cost rather than a contested allocation.
What Procurement Can See Through Showback Data
Showback and chargeback reports are useful for more than engineering behavior change. Procurement teams use cost attribution data to verify that negotiated discounts and committed spend structures are functioning as intended.
If a showback report reveals that a team's AI workloads are consuming inference through an API endpoint not covered by the enterprise agreement, procurement has an actionable signal. The spend is visible and it is outside the commitment structure which means it is not drawing down against a MACC (Microsoft Azure Consumption Commitment) or PPA (Private Pricing Agreement), and it may be paying list price where discounted rates are available.
Mavvrik's GPU Chargeback tracks consumption at the model, team, or tenant level across cloud, on-premises, and hybrid environments, automating attribution through direct integrations with Kubernetes and GPU sharing technologies rather than relying on manual log reconciliation or spreadsheet modeling. (Mavvrik)
For procurement teams, that attribution layer answers a question that cloud-native billing tools cannot: whether the infrastructure terms negotiated are being fully utilized by the teams they were designed to serve, and where spend is flowing outside the committed structure. Discount leakage is one of the most common findings when procurement teams first get visibility into attributed AI spend.
What a Useful AI Chargeback Report Contains
The difference between a showback report that changes behavior and one that generates questions is specificity. A report that tells a product team they consumed $80,000 in AI infrastructure last month leaves them without information to act on. A report that shows inference costs by model, context tier, and feature surface, with a comparison against the previous period and a breakdown of committed versus on-demand usage, gives both engineering and finance something to reason from.
The unit metric that makes AI attribution reports actionable is cost per inference call, adjusted for model and context length. CloudZero defines this as the foundational AI unit economic metric: total cost of a model response divided by inference call volume in a given period, adjusted for model and context. (CloudZero) When product teams can see that one feature costs $0.04 per inference call and another costs $0.40 per call at similar usage volumes, the architectural and procurement conversations that follow are grounded in a number rather than an estimate.
That level of specificity is worth building toward deliberately, starting with the attribution infrastructure during the showback phase, well before the first chargeback invoice goes out.
How Jennifer Aniston’s LolaVie brand grew sales 40% with CTV ads
The DTC beauty category is crowded. To break through, Jennifer Aniston’s brand LolaVie, worked with Roku Ads Manager to easily set up, test, and optimize CTV ad creatives. The campaign helped drive a big lift in sales and customer growth, helping LolaVie break through in the crowded beauty category.
RESOURCES
The Burn-Down Bulletin: More Things to Know
Mavvrik: The Differences Between Chargeback and Showback in FinOps: Why You Need Both A clear breakdown of how chargeback and showback serve different purposes in cloud cost governance, and why running both in sequence produces better outcomes than choosing one.
ProsperOps: IT Showback in FinOps: A Practical Guide Walks through how showback works when centralized Reserved Instance and Savings Plan portfolios are involved, and how savings need to be distributed fairly across consuming teams rather than sitting with the purchasing account.
Logiciel: Showback or Chargeback: What's Working for Engineering Accountability Looks at how teams are applying showback and chargeback to AI workloads, with examples of hybrid models that use showback for R&D and chargeback for production environments.
AWS: Key re:Invent 2025 Launches to Transform Your FinOps Practices Documents AWS's latest capabilities for showback and chargeback in shared EKS clusters, including Kubernetes label import as cost allocation tags and granular split cost allocation for container workloads.
That’s all for this week. See you next Tuesday!



