Blink UX

Hardware Usability Benchmarking UX Research

UX Research: Benchmarking Study for a Fortune-500 Client

UX Research: Benchmarking Study for a Fortune-500 Client

Elevating user experience with wearable devices through series of benchmarking usability tests.

MY ROLE

UX Researcher

Testing Session Moderator

2 Months Total

5 UX Researchers

Qualtrics, Figma, Feedback Panel, Slack

1 Project Manager

1 Director of UX Researcher

2 Research Assistants

May-June & Nov-Dec 2024

DURATION

TEAM

TOOLS

  1. OVERVIEW

I was hired through Blink UX to consult for a Fortune 500 Company leading the industry in wearable technologies. Blink UX is a San Diego local design consulting agency who has had clients including but not limited to Microsoft, NASA, Amazon, Google & Meta. As a UX Researcher at Blink, I was involved in a total of 3 projects from May~December 2024.





Project Type

Benchmarking Study for a Leading Fortune 500 Company

I contributed to the first in a series of benchmarking studies designed to span over five years, with the goal of continuously enhancing the user experience of the client’s products.


Each benchmarking study builds on the previous one, employing consistent research methodologies and analyzing data to track progress. This approach enables the identification of both improvements and areas that require further attention, providing actionable insights to refine the product's performance and user satisfaction over time.

How did we do our Benchmarking Study?
Let’s look at a fun example: Building a Snowman!

Project Overview

Improving UX in Innovative Virtual Reality Wearable Devices

We are tasked to measure specific functionalities and features in the client’s product. Through our research, the clients has some clear goals and deliverables laid out for us.

Goals

Identify pain points and bugs in the technology and product.

Understand the underlying sentiment correlation with frustrations and pain points identified in the study.

Results: Collected and synthesized a total of 150,000 datasets across three projects to provide foundation of the benchmarking study.


Insights Delivered:

Correlation Between User Pain Points & Emotion Scales: Measured user frustration, satisfaction, and frustration levels across product interactions.

Effectiveness of Product Feature Changes: highlight feature changes that improved user experience and vice versa.

Feature Performance Comparisons: Compare older and newer version of features, showing clear trends of UX improvements and declines

Popularity Metrics: Generate prioritized view of user preference based on most and least product features.

Deliverables

Report with qualitative and quantitative
findings

Prioritized task issues list

A cleaned data set

Client Stakeholders

Research Director & Lead Researcher

Research Timeline

The Research Timeline for Benchmarking Studies

We are tasked to measure specific functionalities and features in the client’s product. Through our research, the clients has some clear goals and deliverables laid out for us.

Timeline

Planning

Training

Data Collection

Data Synthesis

(5 Days)

(2 Days)

(5-7 Days)

(5-10 Days)

Communicate with stakeholders, Finalizing on tasks to measure

Creating Qualtrics, Finalizing task flow

Writing up lab notes/ research guidelines

Run pilot testings with stakeholders and participant

Ran 4 two-hour-long usability testings per day, over a span of a month

Compile the Issues List for data synthesis

Identify trend and patterns for product usability and user frustrations

Generate data comparison reports in relation to previous data benchmarking projects

  1. My Role

Planning Out The Testing Task Flow

Series of tasks are going to be tested during the sessions, so we need before we actually start the participant testings, we want to make sure everything is laid out in a logically chronological order.

That required a lot of user journey mapping pre-sessions, and accompanied by our internal pilot testings and external pilot testings.


Process of Planning Testing Flow

Planning Task --> Predicted User Journey Mapping --> Internal Pilot Testing -->
1x External Pilot Testing

Collecting User Inputs and Data

As a research moderator, I conducted 1-on-1 usability studies in a lab setting. During each session, I used a Qualtrics data collection form, followed research procedure guidelines, and observed participants through a computer monitor to analyze their interactions with the product.


Key Observations:

User Journey Flow: Tracking how users navigate and complete tasks.

Disruptive Activities: Identifying bugs, user confusion, or unexpected behaviors.

Real-Time Documentation: Logging findings in Qualtrics to ensure data integrity and prevent misleading insights.

Data Documentation & Synthesis

We collect over 150,000 datasets per study, which is a large amount that can easily become really complicated and confusing when we were trying to make sense of the data.

That is why we decide to dissect data documentation and synthesis into smaller digestible dosage. After every session that we ran, we document hardware issues that we encountered into the Issues List.


Compiling the Issues List: Issue Tracking & Reporting

Logging Device Issues: Recording when and where problems occur and how they impact the data.

Purpose:

Providing quick reports to clients to request bug fixes for device hardware.

Ensuring usability reports reflect only UX-related insights, not hardware failures.

Preparing for Figma Affinity Mapping by distinguishing UX pain points from technical glitches.

This structured approach allowed us to focus on design-driven usability improvements, ensuring a clear separation between UX challenges and hardware malfunctions.


Identifying UX Themes: Figma Affinity Mapping

Using data collected in Qualtrics and insights from the Issue List, we compiled a comprehensive Figma Affinity Map to identify recurring themes and patterns in preparation for the client data report.

Purpose:

Transform low-level user issues into grouped insights.

Identify high-level UX themes that highlight key pain points.

Provide clients with actionable steps to improve the user experience.

This process helped distill raw usability data into clear, strategic recommendations for product enhancement.

Issues Tracking List

Identify recurring themes and patterns, looking at only UX product design issues


Purpose:

Low-level user issues-->Grouped Insights

Prepare for final client data benchmark report. Recommend actionable steps to improve product

Records the location and timing of device’s hardware bugs and issues


Purpose:

Pinpoint outlier data caused by
device hardware bugs

Easy to traceback during affinity mapping creation process

Affinity Mapping

Data Reporting - Client Presentation & Benchmarking

The fact that the benchmarking project’s purpose is to compare with previous projects that had used the similar testing methods and testing metrics on similar tasks, there needed to be sufficient explanation and data evidence backing up drops and ups of task performances.


Our team lead is in charge of directly communicating with the client stakeholders reporting the results of our testings. Whenever they need more clarifications on the number dipping or rising, I am tasked to provide evidence-backed explanation on why a certain task faced significantly lower scores compared to the previous round of testing.


On the report, the performance of each task is indicated through color coding: Green (good), Orange (moderate), Red (poor) for quick readability.

  1. Impact of Our Work

Establishing for 5-Year-Benchmarking Projects

Our projects serve as the foundation for a five-year initiative focused on data collection and usability testing, tracking the product’s evolution through ongoing updates. As the first phase of this long-term benchmarking effort, our work establishes a critical framework by identifying technical challenges, refining testing methodologies, and ensuring the reliability of performance metrics for future evaluations.

Identified

Technical

Challenges

Refined

Testing Methodologies

Ensured

Reliability of Performance Metrics

Our Work Marked The Start of a 5-Year Benchmarking Initiative

Impact:


For the next five years, the client will leverage our project's data to drive product changes and new feature development, guaranteeing a consistent approach across their projects.