AI is showing up everywhere. That visibility can create a false sense of progress. A team runs a few pilots, people use chat tools, and leadership assumes the organisation is "doing AI."

AI is showing up everywhere. That visibility can create a false sense of progress. A team runs a few pilots, people use chat tools, and leadership assumes the organisation is "doing AI."

The evidence is more sobering. In a February 2026 working paper, researchers reported that about 70% of surveyed firms actively use AI, yet over 80% reported no impact on productivity or employment over the prior three years (Yotzov et al., 2026). Separately, McKinsey reported that 72% of organisations used AI in at least one function in early 2024 (McKinsey & Company, 2024). Adoption is high. Measurable impact is harder.

That gap is where "AI maturity" matters. It helps leaders answer a practical question: are we building a repeatable capability, or collecting disconnected experiments.

What AI maturity means for leaders

AI maturity is your organisation’s ability to repeatedly produce business outcomes with AI while managing risk in ways that customers, staff, and regulators can accept. It is less about which tools you own. It is more about whether you can run AI as an operating capability.

Andrew Ng captured the scale of the moment: "I actually have a hard time thinking of an industry that I don’t think AI will transform in the next several years" (Lynch, 2017).

If AI will transform your industry, maturity becomes a competitive differentiator. Measuring it stops you from guessing.

Two lenses that keep the assessment non-technical

You can measure maturity with two leadership lenses.

  • First, value - Are we producing outcomes that matter, and can we repeat them.
  • Second, trust - Are we controlling risk, and can we explain our decisions.

For trust, use established guidance. The NIST AI Risk Management Framework frames risk management as four functions: Govern, Map, Measure, and Manage (Tabassi, 2023). For organisational governance, ISO/IEC 42001 sets requirements for an AI management system so AI work has accountability and continuous improvement, similar to how other management systems operate (International Organization for Standardization, 2023).

For values and stewardship, the OECD AI principles emphasise fairness, transparency, robustness, and accountability (OECD, 2019). You do not need to implement every framework in full to assess maturity. You can use them as anchors for what "good" looks like.

A maturity ladder you can use in an exec conversation

MIT CISR describes four stages of enterprise AI maturity (Weill et al., 2024).

  • Stage 1: Experiment and prepare.
  • Stage 2: Build pilots and capabilities.
  • Stage 3: Develop AI ways of working.
  • Stage 4: Become AI future ready.

This model is useful because it frames maturity as a shift in how the organisation works, not a measure of technical sophistication. MIT CISR reports that financial performance improved at each stage, and that organisations in stages 3 and 4 performed well above industry average in their dataset (Weill et al., 2024).

A leadership scorecard to find gaps fast

Score your organisation across six domains. Use a simple rating for each domain: Red (fragile), Amber (inconsistent), Green (repeatable). Then look for patterns.

1. Strategy and value management

Can you name the few business outcomes you want AI to move this year, with clear owners and targets?

Green looks like a portfolio of use cases tied to measurable outcomes, with clear stop rules when value is not emerging.

2. Data readiness and access

Do teams spend more time finding and cleaning data than improving decisions?

Stage 1 and stage 2 organisations often focus on making data accessible and shareable, so teams can build and validate use cases faster (Weill et al., 2024). If this stays Red, AI becomes slower and more expensive than leaders expect.

3. People and ways of working

Do leaders, managers, and frontline teams understand where AI helps, where it does not, and how to use it responsibly?

MIT CISR highlights AI education and acceptable-use policies as early maturity capabilities (Weill et al., 2024). If training exists only for specialists, adoption tends to rise without consistent business impact, which aligns with the "use without impact" pattern seen in the NBER study (Yotzov et al., 2026).

4. Operating model and delivery discipline

Do you have a repeatable path from idea, to pilot, to scaled rollout?

BCG reports that leaders have scaled more use cases, and that the average financial impact per use case is significantly higher among leading companies (Baltassis et al., 2024). It also warns that companies often set unrealistic maturity ambitions, which can frustrate executives and teams (Baltassis et al., 2024). A clear operating model is how you set realistic timelines and manage expectations.

5. Governance, risk, and accountability

Can you explain who approves AI use, what data is allowed, and how you monitor outcomes and harm?

NIST’s "Govern" function focuses on roles, policies, and oversight (Tabassi, 2023). ISO/IEC 42001 reinforces that AI needs management-system discipline, with clear responsibilities and continuous improvement (International Organization for Standardization, 2023).

6. Measurement and learning loops

Can you show evidence of impact, and can you learn from failures without hiding them?

If you cannot measure outcomes, you cannot mature. Outcome measures are also the fastest way to separate meaningful progress from activity.

Here is an example scorecard you can copy into a leadership pack.

Domain Key
question
Red
(fragile)
Amber
(inconsistent)
Green
(repeatable)
Evidence you
can ask for
Strategy and value management Are we clear on the outcomes AI must move this year Lots of ideas; no measurable outcomes; no owner Some outcomes; mixed ownership; priorities change often Clear outcomes; owners; targets; stop rules A short portfolio list with owners, KPIs, and decision dates
Data readiness and access Can teams reliably access usable data for priority use cases Data is hard to find; quality issues are common; access is slow Data exists for some areas; access depends on relationships Data is discoverable; governed access is routine; quality is measured Data catalog or inventory; defined data owners; data quality measures
People and ways of working Do leaders and teams know how to use AI responsibly in their work Skills sit with a few people; no shared standards Training exists; adoption varies by team; practices differ Role-based guidance; shared practices; learning loops Training completion by role; playbooks; examples of good use and anti-patterns
Operating model and delivery discipline Can we move from idea to pilot to scale predictably Pilots happen; scaling is rare; delivery is ad hoc Some repeatable steps; scaling is slow; handoffs are unclear Clear delivery path; reusable patterns; scaling is normal A standard intake process; delivery checklists; post-implementation reviews
Governance, risk, and accountability Can we explain who approves AI use and how we manage risk No clear approvals; unclear data rules; limited monitoring Some policies; exceptions handled informally; monitoring is partial Clear decision rights; policies; monitoring; audit trail RACI for approvals; policy set; risk register; monitoring reports
Measurement and learning loops Can we show outcomes and learn fast when things fail Activity measures only; outcomes not tracked Outcomes tracked for some use cases; inconsistent reporting Outcome metrics are standard; learnings feed back into process Before/after measures; dashboards; documented learnings and changes

If most domains are Green, you are in a good space. If several are Amber, focus on consistency and repeatability. If any of the foundation domains are Red (data, governance, measurement), address those first because they limit both scale and trust.

How to interpret your results

If you are mostly Green in stage 1 or stage 2 behaviours, you are in a good space if you can prove two things:

  • You can point to specific outcomes that improved.
  • You have clear guardrails that prevent unmanaged risk.

If you are mostly Amber, the issue is often consistency. Many organisations have pilots and tools, but no repeatable ways of working. MIT CISR describes this shift as the move into stage 3, where scaling and reuse become normal (Weill et al., 2024).

If you are Red in governance, data readiness, or measurement, treat those as priority work. Without them, scaling is either slow or unsafe.

Turning the maturity score into a leadership playbook

A maturity assessment only helps if it changes decisions. The hard part is moving from pilots to scaled ways of working. MIT CISR’s 2025 update reports that the greatest financial impact shows up when enterprises progress from stage 2 (pilots and capabilities) to stage 3 (scaled AI ways of working). It argues that this transition needs a united top leadership team, and it frames the work as four challenges: strategy, systems, synchronization, and stewardship (Woerner et al., 2025).

You can use those four challenges as the "what we do next" for your scorecard.

  1. Strategy: Turn "AI projects" into a small portfolio of outcomes. Name owners. Define the metric that proves value. Decide what you will stop. If you cannot stop anything, you are not managing a portfolio.
  2. Systems: Remove the friction that keeps teams trapped in one-off solutions. Aim for shared data access, reusable components, and a clear path to reuse what works. This is how you make success repeatable.
  3. Synchronization: Redesign work so people know where AI fits and where it does not. Clarify decision rights, review expectations, and training by role. Align incentives so teams optimise for outcomes, not activity.
  4. Stewardship: Treat trust as a design constraint, not a policy document. Use a consistent risk management approach across the AI lifecycle. The NIST AI RMF Playbook offers practical, voluntary actions aligned to the AI RMF functions, and it is designed to be tailored rather than followed as a checklist (National Institute of Standards and Technology, 2025).

A practical way to operationalise this is to run your maturity scorecard as a standing leadership agenda. Review it on a regular cadence, and commit to improving one Red domain at a time. That creates momentum, and it keeps maturity grounded in value and trust.

Measuring AI maturity is not about judgement. It is about making progress visible, and turning AI from isolated activity into a capability the organisation can sustain.

References

More articles from Linum Labs