UX Case Study: UX Scoring

UX Scoring

I led a self-initiated effort at Google to build a design quality system—process + tools—that turned subjective debates into objective, evidence-based launch decisions.

Lead Product Designer at Google • 1.5 Years • Org-wide adoption

Hero mockup of the UX Scoring dashboard
TL;DR

Problem: Launches were delayed by subjective debates and inconsistent UX quality, creating friction and a need for a predictable approval path.

Approach: A cross-functional sprint produced a system of standardized guidelines, a scorecard for objective scoring, and auto-linked feedback reports.

Outcome: The system led to predictable approvals, less friction, and higher product quality. It achieved org-wide adoption and measurable business lift.

My Role

Self-Initiated Project

  • Identified the quality gap and proposed a system to leadership.
  • Presented to Directors/VPs and secured sponsorship to proceed.

Design & Planning

  • Partnered with my manager and a Director on goals, deliverables, and timeline.
  • Co-designed with cross-functional designers to shape the final system.

Team Leadership

  • Planned and facilitated a 5-day design sprint.
  • Led weekly rituals across 3 PMs, 2 DEVs, 1 UXR, and 13 UXDs.
Design sprint collaboration photo

5-day sprint to align on approach and validate risks.

Product Overview

UX Scoring is a governance system that evaluates design quality before launch. It creates a shared, objective language for “product excellence” and streamlines go/no-go decisions.

High-level overview of the UX Scoring product

Process + tools: guidelines → scorecard → feedback report.

The Challenge

Products were launching with inconsistent UX and avoidable issues. There were no standard guidelines or metrics, creating friction between teams, last-minute escalations, and costly delays.

I proposed a system to replace subjective debates with objective evidence. With Director/VP support, I formed a cross-functional group to design and pilot the approach.

Diagram: lack of metrics leads to poor UX outcomes

Without shared criteria, “quality” becomes opinion and slows launches.

Diagnosing the Pain Points

I led interviews across UX, PM, Eng, and TLs. The recurring theme: teams needed evidence, not opinions, and earlier signals—not surprises at launch review.

Personas

Key personas for the system

Pain Points

Pain points across roles
“To stop a launch over low design quality, we need evidence, not opinions.” — Nathan, Technical Lead
“Launch approvals are frustrating—I don’t know where we stand until the meeting.” — Luna, Software Engineer
“If design reviews happened earlier, we’d save time and energy.” — Pasha, Product Manager

Defining Our Direction

With leadership, I framed a clear vision and operating principles to guide the build and adoption.

Vision

All products meet product-excellence standards prior to launch.

Mission

Equip teams with tools and guidelines to measure UX quality objectively.

Goals

Create UX Guidelines

Define consistent attributes and patterns to avoid repeat issues.

Design Measurement Tools

Provide a scorecard to evaluate UX and surface targeted feedback.

Generate UX Evidence

Attach scores and guideline links to justify decisions.

Guiding Principles

Simple to Use

Reduce—not add—friction between teams.

Open Access Data

Scores and launch data are transparent by default.

Scalable

Works across product areas and team sizes.

Designing an Ecosystem for Quality

Research made it clear: one tool wouldn’t fix a process problem. I led the design of a system that teams could actually run end-to-end.

1) Standardized Guidelines

A centralized, practical set of visual and usability standards gave teams a shared language—and prevented repeat issues.

Visual and usability guidelines examples

A single source of truth for UX quality expectations.

2) Data-Driven Scorecard

Designers evaluate a product against the guidelines to produce a score. This shifted reviews from opinion to evidence.

UX Scorecard interface

Scorecard turns subjective feedback into objective criteria.

3) Linked Feedback Reports

Automated reports attach evidence with links to the exact guidelines and give clear, prioritized next steps.

Comprehensive feedback report

Actionable guidance that teams can ship against.

Product Design & Evolution

From early wireframes to high-fidelity, we iterated during a 5-day sprint and weekly critique with 13 designers to remove friction from every step.

UX Score Request Form

A lightweight intake that captures scope, stage, and owner in under a minute—auto-routing requests to the right reviewer and setting clear SLA expectations.

Score request form

Scorecard

Evaluator view maps each criterion to a guideline, applies weights, and flags blockers with rationale and examples—turning debate into a defendable score.

Scorecard interface

UX Scores Repository

Central, searchable history of reviews by product and release—helping teams compare trends over time and reuse prior feedback instead of reinventing it.

Repository of UX scores

Design Evolution

Snapshots from low-fi flows to polished UI show how sprint insights simplified tasks, removed redundant steps, and aligned terminology across teams.

Design evolution from low-fi to high-fi

The Impact: A New Standard for Quality

UX Scoring streamlined launch reviews, reduced friction between design and engineering, and raised product quality across teams.

Process Improvement

Clear, predictable approvals replaced last-minute surprises and escalations.

Improved launch process diagram

Approval flow clarity = fewer delays.

Business Results

Higher-quality launches correlated with stronger performance and satisfaction.

Chart showing correlation between scored projects and revenue lift

Quality up → outcomes up. Simple.

Retrospective & Lessons Learned

The Power of Phased Rollouts

Piloting with a small group let us learn fast, refine criteria, and build advocates before org-wide adoption.

Standardization vs. Autonomy

Standards set the bar; flexible application unlocked adoption and ownership across teams.

Proving Value with Data

Tracking outcomes post-launch sustained executive support and funding.

Roadmap

Next steps to scale UX Scoring over 12 months.

0–6 Months

  • Integrate score requests into PRDs / issue trackers.
  • Add role-based templates (PM / Eng / UX).
  • Publish a clear “quality bar” with pass/fail examples.

6–12 Months

  • Automate basic checks (contrast, tap targets, layout).
  • Attach lint output + screenshot diffs to reports.
  • Self-serve dashboard for score trends and rework.
Metrics to Watch

Time-to-approval ↓

Rework after launch ↓

Reviewer coverage & adoption ↑