How to Measure AI Product Impact: The Bullseye Framework for Power, Speed, Impact, and Joy

Most AI dashboards measure usage. They should measure outcomes. Here's the four-pillar framework and the three-layer telemetry stack that replaces vanity metrics.

A
Arpy Dragffy · · 7 min read
Editorial photograph: How to Measure AI Product Impact: The Bullseye Framework for Power, Speed, Impact, and Joy
Photo: Generated via Flux 1.1 Pro
Overview
  • Most AI product teams measure power (what it can do) while ignoring impact (what outcomes change) and joy (whether users trust it enough to come back).
  • The Bullseye framework calibrates four pillars — power, speed, impact, and joy — because optimizing for one while ignoring the others produces products that demo well and fail in production.
  • Impact blindness — the inability to see whether AI is helping or harming — is the defining measurement crisis of the agent era.
  • A three-layer telemetry stack (binary + outcome + satisfaction) replaces the single-layer dashboards that most teams use today.

Why are AI product metrics lying?

Most AI product dashboards track adoption: tasks completed, messages sent, workflows automated, time saved. By those metrics, enterprise AI deployments are succeeding at record rates.

The metrics are lying. They are measuring activity, not impact. And in the agent era — where AI systems act in the background without visible user interaction — the gap between what dashboards report and what actually happens is widening into a crisis.

On Episode 1 of the Product Impact Podcast, we introduced the framework we use at PH1 Research to diagnose this problem: the Bullseye.

What is the Bullseye framework?

The Bullseye is a four-pillar calibration model for AI product quality. It is not a scorecard. It is a calibration — the recognition that optimizing for one pillar while ignoring the others produces products that demo well and fail in production.

Power — what the product can do. What tasks can it complete? Across what constraints? What is the ceiling of its capability? Power is what gets funded. It is also table stakes in 2026, because every product runs on the same foundation models.

Speed — how quickly value is produced. Not just latency (does it respond fast?) but time to completion, time to confidence, and time to learning. An AI that responds in 200ms but takes 15 minutes of prompt iteration to produce a usable output is fast in the wrong dimension.

Impact — what outcomes actually change. Did the user accomplish their goal? Did the business metric move? Did the outcome stay moved, or did it revert? Impact is where most dashboards fail, because they measure activity (the user interacted with the AI) and call it impact (the AI helped).

Joy — whether users feel confident, in control, and willing to come back. Joy is not satisfaction surveys. It is confidence, clarity, control, and willingness to delegate again. If users don't trust the AI enough to hand it a second task, the product has a joy problem that no amount of capability improvement will fix.

The F1 analogy is instructive: the car with the biggest engine does not win. Weight undermines success and adds danger. What wins races is calibration — the setup, the tires, the balance, the strategy for that track, that condition, that driver. Products are the same. Now that every product has a turbo engine (frontier models), the competitive edge is calibration across all four pillars, not raw power.

What is impact blindness?

Impact blindness is the inability to see whether AI is actually helping or harming. It is the defining measurement crisis of the agent era.

Historically, UX created visible signals: hesitation, abandonment, retry patterns, support tickets. These signals told product teams when something was wrong. With AI agents operating in the background — booking flights, closing tickets, updating CRMs, drafting emails — those signals compress or vanish. The system acts without the user watching. The user sees the output but not the process.

The result: your system logs a successful completion. The user never returns. You don't know why until churn shows up months later — if you notice at all.

Example: AI travel booking. An agent books a flight. Success logged. But it picked a brutal layover, ignored seat constraints, and the user doesn't notice until the confirmation email. They don't complain. They think, "I'm never using this again." That is a silent impact failure that no binary completion metric can detect.

Example: AI support resolution. A ticket is created, the agent resolves it, the resolution rate looks great. But the user felt blamed, had to double-check everything manually, and stopped trusting the product. The dashboard shows a green checkmark. The user is gone.

Success does not mean satisfaction. A completion event can hide frustration, regret, distrust, or a feeling of lost control.

What should AI product teams measure instead?

A three-layer telemetry stack replaces the single-layer dashboards most teams use:

Layer 1: Binary telemetry (what happened)

Completed, failed, error types, retry count, escalation to human. This is necessary but not sufficient. Binary telemetry tells you the car crossed the finish line. It doesn't tell you whether the driver wants to race again.

Layer 2: Outcome telemetry (did it actually help?)

  • Time to completion — not just response latency, but end-to-end time from intent to outcome
  • Rework rate — how often does the user correct, undo, or redo the AI's output?
  • Ghost abandonment — completion followed by non-return. The user got an answer and never came back. Was that success or silent failure?
  • Repeat delegation rate — did they hand the AI a second task? This is the strongest signal of real impact. A user who delegates once and never returns is telling you the AI failed, regardless of what the completion log says.

Layer 3: Satisfaction telemetry (how did it feel?)

  • Post-task confidence — how confident was the user that the AI's output was correct?
  • Perceived control — could they steer, stop, undo? Did they feel in control of the process?
  • Trust delta — are they more or less willing to delegate to the AI next time? This is the leading indicator that predicts whether adoption holds or decays.

If you can't capture Layer 3, you can't tune for joy. And if you can't tune for joy, you lose repeat usage — which means your adoption metrics will look healthy until the moment they collapse.

What's wrong with LLM-as-judge evals?

AI evals — the automated quality assessment tools that most teams rely on — use a strong model to score outputs against a rubric. The LLM-as-judge framework has become the standard for evaluation at scale. It is useful. It is also systematically biased.

Three documented biases affect LLM-as-judge reliability:

Position bias — the model tends to prefer whatever appears in the first position, regardless of quality.

Verbosity bias — longer responses are rated as more complete and more correct, even when they contain more filler.

Self-enhancement bias — the model rates LLM-produced responses as more accurate than human-produced responses.

LLM-as-judge can tell you if the output became more rubric-compliant. It cannot tell you if you improved joy, trust, or willingness to delegate. It is one rung in an evaluation ladder — not the ladder. The ladder requires all three telemetry layers, with human-in-the-loop validation for the dimensions that automated judges systematically miss.

How should product teams apply this?

Start with the question the Bullseye forces: which pillar is weakest, and is the team even measuring it?

Most teams in 2026 are measuring power (capability benchmarks) and speed (latency). Very few are measuring impact (outcome change) or joy (trust and willingness to return). The teams that ship more activity while outcomes degrade are the ones that will face the sharpest corrections.

The Bullseye framework and the three-layer telemetry stack are the tools PH1 Research uses with product teams to diagnose this gap. AI Value Acceleration applies the same lens to enterprise deployments — finding where the AI is active but not valuable, where dashboards show green while users disengage.


Listen: Product Impact Podcast S02E01 — Why Your AI Metrics Are Lying to You

Related:
- Gartner Says 40% of Agentic AI Projects Will Fail — the cascading error problem
- What AI Does to Human Thinking — measuring cognitive sovereignty
- Enterprise Context Is the AI Moat — knowledge-first approach

Sources:
- Product Impact Podcast S02E01 — primary source for framework and examples
- PH1 Research
- AI Value Acceleration

A
Arpy Dragffy

Founder, PH1 Research · Co-host, Product Impact Podcast

View all articles →

Hosted by Arpy Dragffy and Brittany Hobbs. Arpy runs PH1 Research, a product adoption research firm, and leads AI Value Acceleration, enterprise AI consulting.

Get AI product impact news weekly

Subscribe

Latest Episodes

All episodes
7: $490 Billion in AI Spend Is Delivering Nothing — Orchestration Is the Fix
EP 7

7: $490 Billion in AI Spend Is Delivering Nothing — Orchestration Is the Fix

Apr 17, 2026
6. Robert Brunner Was the Secret to Beats' & Apple's Success — Now He's Redefining AI for the Physical World
EP 6

6. Robert Brunner Was the Secret to Beats' & Apple's Success — Now He's Redefining AI for the Physical World

Apr 9, 2026
5. The Human Impact of AI We Need to Measure [Helen & Dave Edwards]
EP 5

5. The Human Impact of AI We Need to Measure [Helen & Dave Edwards]

Mar 30, 2026
4. The AI Agent Era Will Change How We Work
EP 4

4. The AI Agent Era Will Change How We Work

Mar 19, 2026
3. Win The AI Context Wars — Unlock The Value of Data [Juan Sequeda ]
EP 3

3. Win The AI Context Wars — Unlock The Value of Data [Juan Sequeda ]

Mar 12, 2026

Related

6