Sam Copsey
← Back to blog

Measuring What Matters: A Framework for AI Success Measurement

Download the framework (PDF)

A 2025 MIT study found that 95% of generative AI projects fail to demonstrate ROI. Not because the projects are failing. Because the measurement is wrong.

Most teams measure AI success the same way they measure traditional software: hours saved, cost per item, velocity. These metrics tell you how much you built and how fast. They tell you almost nothing about whether what you built is actually working.

I ran into this problem firsthand. Our engineering team ships 10+ concurrent AI and automation products across document review, vendor research, service monitoring, and cloud management. When leadership asked “is this stuff making a difference?”, hours saved alone could not answer the question. A document reviewer that saves 4 hours per review but produces outputs nobody trusts has not saved 4 hours. A research platform that creates entirely new capability for the sales team shows zero hours saved, but delivers significant value.

So I took what we’d learnt and decided to write a framework.

The thesis

Hours saved is necessary but insufficient. Organisations that achieve sustainable ROI from AI investments use 6 to 10 complementary KPIs spanning multiple dimensions. High maturity organisations measure across 76% of their defined KPI categories, compared to 25% for low maturity ones.

Qualitative metrics like satisfaction, trust, and cognitive load appear weeks after deployment. Hard financial ROI typically takes 12 to 18 months. The soft signals predict whether the hard ROI will follow. Products with high early adoption and satisfaction almost always deliver financial returns. Products with low adoption and satisfaction almost never do, regardless of their theoretical time saving potential.

Seven dimensions of success

The framework synthesises research from SPACE (Microsoft Research), DORA (Google), Balanced Scorecard, Gartner, NIST, and MITRE into seven measurable dimensions:

  1. Delivery Performance. Are we shipping reliably and predictably?
  2. Product Quality & Reliability. Does what we ship work correctly and consistently?
  3. User Adoption & Engagement. Are people actually using what we built?
  4. User Satisfaction & Experience. Do users find the product useful and usable?
  5. Trust & Confidence. Do users trust AI outputs enough to act on them?
  6. Cognitive Load & Developer Experience. Does the product reduce or increase mental burden?
  7. Strategic & Business Impact. Is this driving outcomes that matter to the business?

No single dimension tells the full story. Together they show what was built, how well it works, whether people use it, how they feel about it, what it costs them cognitively, and what strategic value it creates.

Value moments

The framework introduces the concept of value moments, the specific user interactions where a product delivers its intended value. Not “dashboard viewed” but “incident detected before SLA breach”. Not “document submitted” but “review completed and accepted”. Not “report generated” but “research influenced a decision”.

Hours saved should be calculated per value moment, not per login. Satisfaction surveys should trigger at value moments, not at random. Adoption should track value moment frequency, not just access counts. This single shift in measurement focus changes everything about what you see.

What is in the whitepaper

The full document covers:

  • Why quantitative only measurement fails, with evidence from MIT, Deloitte, and DORA
  • A complete per dimension deep dive with KPIs, scoring, and benchmarks
  • Validated psychometric instruments you can use directly: SUS, S-TIAS, NASA-TLX, SPACE, CES, eNPS
  • Guidance on measurement at small scale (because most internal teams have 5 to 15 users, not hundreds)
  • A three month implementation roadmap: baseline, instrument, report
  • Per product measurement plans and quarterly scorecard templates
  • Responsible AI as a cross cutting lens across all seven dimensions

Who this is for

If you are an engineering manager, product lead, or anyone trying to answer the question “is our AI investment actually working?”, this framework gives you the instruments and structure to answer it properly. It was built for a real team shipping real products. Every metric has a specific measurement method, a scoring methodology, benchmarks where available, and a recommended cadence.

The framework is designed to be adapted. Replace product names with your own, adjust the cadences to fit your team’s size, and start with the dimensions that matter most for your context. The goal is not to measure everything. It is to measure the right things.


Want the survey starter pack?

I have put together a ready to use set of survey instruments based on this framework, including SUS, trust, and cognitive load questionnaires adapted for internal AI products. If you would like a copy, or want to chat about how to apply this in your organisation, drop me a line at sam@samcopsey.co.uk.