How to measure the ROI of your AI projects
Moving beyond the hype to actual business impact
How to measure the ROI of your AI projects
Every AI project I have been involved in over the past two years has had the same conversation at some point. Leadership asks what the return on investment is and the team struggles to give a clear answer. Sometimes its because the measurement was not planned from the start. Sometimes its because the team measured the wrong things.
I want to share the framework I use to measure AI project impact in a way that actually resonates with business stakeholders.
Start with the baseline before you build
This sounds obvious but it is the step that gets skipped most often. Before you build anything, measure how the process works today.
For each process you are targeting with AI, capture:
- How long does it take a human to complete this task
- How many people are involved
- What is the error or rework rate
- What does it cost (time x loaded hourly rate)
- What is the throughput - how many tasks are processed per day/week
If you dont have this baseline, you have nothing to compare to after deployment. I have seen teams claim huge wins but when you dig into it there is no pre-deployment data to validate the claim.
The metrics that matter
Time savings is the most straightforward. If a document review that took 45 minutes now takes 5 minutes with AI assistance, that is 40 minutes saved per document. Multiply by volume and the hourly cost of the person doing it and you have a number you can take to the CFO.
Quality improvement is harder to measure but often more important. Error rates, rework rates, customer complaints, compliance incidents - these are the quality signals that AI can move. The problem is they require good measurement infrastructure that many organisations dont have.
Throughput increase is relevant when AI removes a bottleneck. If your legal team can review 10 contracts a week but demand is 25, AI assistance that gets them to 20 reviews is a meaningful capacity improvement even if it doesnt directly reduce headcount.
Deflection rate applies to customer-facing applications. What percentage of queries does the AI handle without escalating to a human? Track this alongside customer satisfaction - a high deflection rate with low satisfaction is not a win.
A simple tracking template
Here is the structure I use for tracking impact across a project:
Process: [Name of process]
Owner: [Business owner]
Baseline (pre-AI):
- Average time per task: X minutes
- Volume per month: X tasks
- Error rate: X%
- Monthly cost: $X (time x hourly rate x volume)
Post-deployment (measure monthly):
- Average time per task: X minutes
- Volume per month: X tasks
- Error rate: X%
- Monthly cost: $X
- AI infrastructure cost: $X
Net monthly saving: $X
Cumulative saving: $X
Project cost to date: $X
Payback period: X months
Keep this simple and update it monthly. Share it with stakeholders regularly so the value of the project is visible and not just implied.
The costs that get forgotten
Teams often undercount the cost side of the equation.
- Inference costs - your monthly Azure OpenAI bill will grow with usage, build this into your projections
- Maintenance - models change, prompts need tuning, integrations need updating
- Human oversight - if your process requires human review of AI outputs, count that time
- Training and change management - getting users comfortable with the new process takes time
- Data preparation - cleaning and structuring data for your AI application is real work
A project that saves 100 hours a month of manual work but requires 20 hours a month of maintenance and monitoring has a net saving of 80 hours. Still good, but be honest about the numbers.
What good looks like
The best AI projects I have seen have a few things in common:
- They started with a specific, measurable problem rather than "lets use AI"
- They measured baseline performance before building anything
- They launched small, measured the impact, then expanded
- They kept the business stakeholder involved throughout so there were no surprises
The worst ones started with the technology and worked backwards to find a problem. They were usually impressive in demos but struggled to show real impact in production.
AI is genuinely a transformative technology but its still a tool. Use it where it solves a real problem, measure whether it actually does, and be honest about both the wins and the failures. That is how you build the credibility to get the next project approved.