How have you calculated ROI of AI solutions (including agents) that you've rolled out at your firm? Are there specific KPI's that you've focused on and how have you measured (and validated) the value?
Sort by:
We calculate ROI using a combination of quantitative and qualitative factors, which vary depending on the client. Quantitative factors include utilization, time saved, cost savings, and new revenue opportunities. Qualitative factors include client satisfaction, deliverable quality, retention, and engagement. This two-pronged framework provides a validated view of business value, benchmarked over time to guide strategic deployment.
The most effective approach we’ve found—with partners and customers—is to evaluate ROI at the use‑case level, tied to the P&L. We segment by value driver and validation method:
• Efficiency plays (e.g., back-office): hard ROI via FTE-hour reduction, cycle-time compression, error-rate cuts, and throughput;
• Growth plays (e.g., sales force automation at scale): blended ROI via incremental revenue, conversion lift, pipeline velocity, and adoption/usage. Example: we enabled 500+ reps to access actionable insights via voice in WhatsApp and reported improved time-to-insight and call prep, translating into higher contact rates and win rates.
Principles:
• Start with the unit of value (per case, per ticket, per rep) and roll up.
• Isolate impact (control groups, before/after) and de-bias with counterfactuals.
• Track both cash and non-cash benefits (risk, compliance, CX), but prioritize cash conversion within 12 months.
• Define owner, baseline, target, and validation plan before scaling.
In short: ROI is case-by-case, but governed by a consistent framework that ties operational metrics to financial outcomes and uses disciplined measurement to validate value.
"AI" is to broad to provide a standard set of KPIs.
If we are talking specifically about tool performance, these are measured against various benchmarks and are rarely relevant to business-enabling problematics.
If we are talking about business-enabling performance, performance of humans operating AI tools and quality of processes within which that happens need to be factored.
Specifically for ROI calculations, there is a shift of parameters to account for - e.g., if you are using generative AI tools there are licenses or processing volumes, or if you are using agentic AI tools there is cost of external tool calling. In both cases there is also adjustment to the compensation/title/role tripods for humans running these machines. Totals would give you a bang for the buck which you could compare with non-AI baseline.
This is a multi-facted beast. Gartner did some interesting research on it, splitting application of (Gen)AI into stuff that makes current processes run more efficiently (think e.g. MS Copilot, ChatGPT, etc.), stuff that adds new capabilities and changes/extends current business processes to generate more value, and stuff that is transformational and opens up entirely new processes, value streams, business models.
The argument is that financial ROI generally makes sense for the latter two, but less so for the fist category. There, it is mostly employee satisfaction / NPS. So think about that as return-on-employee rather than return-on-investment. The latter two categories are business cases like introducing any other new capability.
I can only echo this. The only thing we did at one point in time was e.g. comparing things that are clearly measurable. So we enrolled entire teams on github Copilot and compared their Software Delivery KPIs overall some months before and some months after. There are clear indicators that teams are more productive (velocity increased from as low as 5% to as high as 50%) as well as code quality slightly decreasing. We also include factors like team happiness into that equation. Long story short: what we were looking for is a reason to justify the expense and that has been found. Cause even if a team only gets a couple of percent more productive there is a strong argument to cover the cost associated.
That makes sense - are you able to point me to the Gartner research you referenced?
We're encouraging clients to estimate ROI in both hard and soft savings compared to cost of deployment and licenses. "Hard" savings being revenue-impacting and direct cost savings (including through reduced headcount), which is what most CFOs care about. "Soft" savings may still have merit if they address particular pain points across the org that affect a large number of users, but we don't want every employee to work on the things that annoy them about their jobs without bringing real dollar value to the company.
In terms of calculation, it's much easier with the hard savings vs. soft savings, but if you're saving folks time in internal processes, it may reduce user downtime or improve customer experience, which can still have some dollar amount attached.