Skip to content
— / Resource

The 174 AI literacy assessment rubric

· 12 min read

The full methodology behind the 10-minute organizational assessment — three dimensions, twelve questions, the scoring logic, and how to interpret your own report honestly.

The assessment most companies offer ends with a sales call. Ours ends with a report you can forward to your CHRO, your COO, or your CEO. To make that report defensible, the rubric behind it has to be public. This is that rubric — every question, every option weight, every bucket threshold, and the logic that turns answers into a recommended rollout shape.

If you’d rather just see your own number, the 10-minute assessment is the fastest way. If you want to read the methodology first — or you’re the kind of buyer who reads methodology before clicking anything — this page is for you.

Why three dimensions

Most AI-literacy assessments measure one of two things: tool usage (“does your team use ChatGPT?”) or sentiment (“how excited are people about AI?”). Both are misleading. Tool usage misses whether the work is any good; sentiment misses whether anything is shipping.

The three dimensions in the 174 rubric — adoption, capability, and governance — exist because they’re the smallest set that captures what actually matters for a mid-market rollout:

  • Adoption is whether AI is in the work, day-to-day, across the people who would benefit. It’s the breadth signal.
  • Capability is whether the work being done with AI is actually good — prompts that get reused, output that gets evaluated, agents that ship. It’s the depth signal.
  • Governance is whether the program can survive scale — policy, quality controls, executive sponsorship. It’s the durability signal.

Drop any one and you can construct an organization that looks healthy on the other two but falls over the moment a rollout actually begins. We considered splitting governance into “policy” and “controls” — they’re meaningfully different — but in practice mid-market buyers think of them together, and conflating them costs us very little. Three dimensions, scored 0–100 each, bucketed into Emerging / Developing / Mature.

The questions

All 12 questions are below. Each option carries a score weight; the total possible score per dimension is the sum of each question’s max-option score. A buyer’s percentage in each dimension is (actual / max) * 100, rounded.

Adoption

A1. What share of your team uses AI tools in their day-to-day work?

Single select. The breadth-of-use signal.

OptionScore
Less than 10%0
10–30%25
30–60%60
More than 60%100

A2. Which AI tools are in active use across your team?

Multi select (up to 6). Each selection contributes — “None of the above” is its own option scored at 0 to keep the math honest. The mix matters: a team running custom agents alongside general assistants is signalling a different maturity than one running only ChatGPT.

OptionScore
ChatGPT / Claude / general assistants20
Copilot for code (GitHub, Cursor, Windsurf)20
AI features inside existing SaaS (Notion AI, Slack AI)15
Custom prompts or GPTs in shared workspaces20
Internal agents / automations (n8n, Zapier AI, custom)25
None of the above0

A3. How often does your team use AI for real work (not experimentation)?

Single select. Frequency-of-real-use, distinguished from playing-with-it.

OptionScore
Rarely or never0
A few times a month25
Weekly60
Daily100

A4. How would you rate your team’s satisfaction with current AI tools?

Single select. Captures whether adoption is durable or about to churn.

OptionScore
Frustrated — outputs aren’t reliable0
Mixed — some wins, some misses33
Generally positive66
Strong — it’s changed how we work100

Capability

C1. How would you rate your team’s prompting skill?

Single select. Honest answer beats aspirational.

OptionScore
Most people just type a question0
Some patterns, but ad-hoc33
Documented prompts that get reused66
Evaluated, versioned, reviewed prompts100

C2. How does your team evaluate AI output before relying on it?

Single select. Eyeballing and structured evaluation are two different rooms.

OptionScore
We don’t — we eyeball it0
Spot checks by a senior person33
Informal rubrics or checklists66
Structured evaluation with documented rubrics100

C3. How often does your team use AI in multi-step / agentic workflows?

Single select. Beyond single prompts — chained, automated, or agent-like.

OptionScore
Never0
Experimenting33
A few production workflows66
It’s a core part of how we work100

C4. How does your team currently learn new AI techniques?

Multi select (up to 5). Captures whether learning is structured or entirely incidental.

OptionScore
YouTube and Twitter / X threads10
Internal Slack channels and word of mouth15
Courses or certifications25
Internal lunch-and-learns or workshops25
A formal AI literacy program25

Governance

G1. Do you have a written AI usage policy?

Single select. The single highest-leverage governance instrument.

OptionScore
No0
Drafting one33
Yes — published but not enforced66
Yes — published, trained, enforced100

G2. Who owns AI enablement in your organization?

Single select. Ownership is the difference between “we should do AI” and “we are doing AI.”

OptionScore
No one in particular0
IT or Security — mostly risk management33
L&D or People Ops66
A dedicated AI / transformation lead100

G3. Are there quality controls for AI output that ships externally?

Single select. Customer-facing copy, code, contracts, decisions.

OptionScore
No controls0
Manager review only33
Documented review steps66
Documented + enforced + audited100

G4. What’s the level of executive sponsorship for AI literacy?

Single select. Without an exec sponsor, the program lives until the next budget review.

OptionScore
No exec sponsor0
Verbal support, no budget33
Budget allocated, no clear program66
Funded program with exec accountability100

How scoring works

Each dimension is computed independently:

  1. Sum the scores for the answers given to the questions in that dimension.
  2. Sum the maximum possible scores for the same questions.
  3. Divide and multiply by 100. Round to the nearest integer.

For Adoption (questions A1 + A2 + A3 + A4), the maximum is 100 + 100 + 100 + 100 = 400. A response earning 0 + 40 + 25 + 33 = 98 produces an Adoption score of round(98 / 400 * 100) = 25.

For Capability and Governance, the same arithmetic applies with their respective maxima.

The overall score is the rounded average of the three dimension scores. There is no weighting — we treat the dimensions as equally important on purpose, because the failure mode of weighting is letting a single strong dimension hide weakness elsewhere.

How buckets work

ScoreBucket
0–39Emerging
40–69Developing
70–100Mature

The thresholds aren’t tuned to a normal distribution — they’re tuned to the rollout decisions a buyer needs to make. A score below 40 in any dimension means the program for that dimension is not yet running. A score between 40 and 69 means the program is running but inconsistent. A score of 70 or higher means the program is durable and ready to scale or deepen.

How recommendations are generated

For every (dimension, bucket) pair, the rubric carries two pieces of copy:

  • The highest-leverage gap. What’s actually limiting progress in that dimension at that maturity. Phrased as a diagnosis, not a complaint.
  • The recommended next move. What 174 would suggest you do first. Phrased as a single concrete action, not a workshop list.

These pairs aren’t generated dynamically — they’re authored. We update the copy when we learn something new from a real engagement; we don’t update them per-buyer. The full mapping is in src/data/assessment.ts in the marketing repo if you want to read every variant.

The overall rollout shape is computed from your overall score plus your weakest dimension:

  • Below 40 overall: a 1–10 seat pilot, in the team most ready to move.
  • 40–69 overall: a department-wide rollout, starting with the weakest dimension.
  • 70+ overall: org-wide concierge — the program scales now.

How to interpret your own score

Three honest principles:

  1. The rubric rewards honest answers. Picking aspirational options (“we’re drafting one” when you haven’t started) inflates the bucket and recommends the wrong rollout shape. The point of the assessment is to see clearly, not to feel good.

  2. Low scores in Emerging are useful, not damning. A 25 on Adoption is a clear instruction: visibility and access first. A 25 on Capability is a clear instruction: curriculum first. The bucket is your starting line, not a grade.

  3. Mature in one dimension, Emerging in another is the most common shape. Mid-market companies routinely have strong adoption and weak governance, or vice versa. The assessment surfaces that asymmetry on purpose — most rollout failures are governance failures dressed up as adoption successes.

If you’ve taken the assessment and want to retake it after a few months of work, the URL is the same. The report you generate is dated, so you can put two reports side by side and see the actual lift.

— / Next move

Where does your org actually stand?

Ten minutes. Three dimensions. A leadership-shareable baseline.