Pioneering a voice-first AI product
MasterClass members want to learn from and connect with the worldβs greatest experts. As AI matured, we saw an opportunity to deliver the personal and practical guidance that members had been asking for: a standalone, voice-first AI product that lets them call an AI version of their favorite expert.
Company: MasterClass
My Role: Product Designer
Project Time: One year
The Challenge
Design a new AI product that lets users have a voice call with the world's greatest experts.
My Role
I worked as the sole product designer on this project. I led all UX research, concept development, and end-to-end design. I collaborated closely with engineering and marketing to bring the MVP to life.
Our beta generated 6,000 conversations with 81% positive sentiment
Understanding the Problem
I conducted competitive research across the AI landscape to identify key opportunities, then led a full-team brainstorm with product, engineering, data, and marketing to align on what our product should be.
COMPETITIVE ANALYSIS
βWhat does the current AI landscape look like, and what can we learn from it?β
I conducted a wide-ranging analysis of 30 AI products, looking at features, pricing, membership structures, voice capabilities, and unique value props. I synthesized everything into key learnings, standout products, and opportunities we could act on.
AI TEAM BRAINSTORM
βWhat is our vision for this product? What are the must-haves?β
I reframed my research learnings as user problems, then facilitated a full-team brainstorm across marketing, leadership, product, engineering, and data to generate ideas around each one. We voted as a team to align on direction, which set me up to move into visual explorations.
HMW create a more human-like AI experience?
HMW create features that stand out?
HMW create a practical tool and a trusted advisor?
Our MVP should:
Make it easy to start a call
Provide guardrails for unpredictable AI models
Highlight AI expert credentials
Early Concept Testing
After defining our MVP, I built a lo-fidelity prototype and ran usability tests to get an early read on the design direction.
LO-FIDELITY PROTOTYPES
βHow should the main mechanics work?β
With our MVP requirements locked in, I jumped into lo-fidelity prototypes to start exploring what this product could look and feel like. The goal here was to pursue as many different directions as possible in order to narrow our vision for usability tests.
USABILITY TESTS
βWhat do users think? Are we headed in the right direction?β
After narrowing down to a single solution, I conducted usability tests on usertesting.com with 12 participants on mobile devices. The goal of these tests was to determine if there were any major usability issues and to discover opportunities for iteration.
βIβm unsure what to ask nextβ
βIβm skeptical that this AI knows what heβs talking about.β
βItβs really frustrating when he interrupts me.β
Building the Solution
After gathering insights from users, I iterated on our early concepts and collaborated with our marketing team to create hi-fidelity designs.
VISUAL EXPLORATIONS
βWhat should OnCall look and feel like?β
I moved into high-fidelity designs and went through several rounds of iteration. I kept user feedback at the center while incorporating input from leadership, product, and marketing to refine and shape the final experience.
Home Page Explorations
Call UI Explorations
The Solution
After iterating on the feedback, I handed off the designs to our engineering team
SOLUTION REVIEW
βDo these designs meet the product requirements and solve the user problem?β
I iterated to the final version based on close collaboration with product stakeholders, leadership, and the Marketing team. I handed the designs off to engineering and was highly involved in the QA process up until the new product shipped.
Make it easy to start a call βοΈ
Users can make a call in a single click
Quick call buttons appear at the top of the screen
Conversation starters throughout the homepage reduce friction to start a call
Guardrails for Unpredictable AI π¦
Call states and user prompts help communicate what the AI is doing and how the user should engage. Call feedback allows users to report if something went wrong.
Highlight AI Expert Credentials π
AI Expert page gives details on why this person is an expert in their field and can be trusted.
Positive User Sentiment
π
Positive User Sentiment π
77% of users rated their call experience as 4 or 5 stars
Easy To Use
π€©
Easy To Use π€©
Next Steps & Reflection
Once we launched the Beta, I provided next steps and reflected on what I would have done differently.
NEXT STEPS
βWhat capabilities and features should we focus on next?β
After launching our beta project and running UXR, we found that users crave texting capabilities. Given our robust library of chefs, we also explored a baking/cooking companion.
Chat & Transcript Explorations
Baking Companion Explorations
REFLECTION
βWhat would I have done differently?β
I would have connected with our engineers earlier in the process in order to better understand the capabilities of our AI experts. The product team thought that the AI would be more advanced than was capable with the time allotted. In the future I would confirm the AI capabilities before jumping into designs.
View Other Projects
MasterClass π Growth Team
MasterClass π Engagement Team

