Product Management 11 guests | 14 insights

Usability Testing

Usability testing is the practice of watching users interact with your product to identify friction, confusion, and unexpected behaviors. It's not 'junior' work; micro-level testing can drive massive business value through small, tactical optimizations. Always present multiple options and never trust your own internal intuition.

The Guide

5 key steps synthesized from 11 experts.

1

Test with users who have zero skin in the game

Recruit prospective users who aren't invested in the product's success to ensure honest, unbiased feedback. Use platforms like UserTesting.com or Voicepanel to automate recruitment. Watch them struggle and listen to them describe their thought process in real-time.

Featured guest perspectives
"We would have an idea in the morning, come up with some sort of functional prototype, recruit a bunch of people that are legitimately good prospective users, but have zero skin in the game, ship fast so people can start playing with it. In the afternoon, we're already running pretty full scale experiment."
— Grant Lee
"It's amazing to me how you can find 10 random people on the internet and they can give such astute feedback that then is so representative for such a large number of people."
— Melanie Perkins
2

Use faked versions to validate before coding

Run 'Wizard of Oz' tests where humans manually perform automated tasks behind a facade. Use 'Fake Door' tests to measure user intent and click-through rates on features that don't exist yet. Progress from 'fish fooding' (team) to 'dogfooding' (company) to alphas to increase confidence iteratively.

Featured guest perspectives
"Initially you fake it, you do a fake door test, you do a smoke test, Wizard of Oz tests. We used a lot of those in the tabbed inbox by the way, one of the first early versions was actually we showed the tabbed inbox working to people. But it wasn't really Gmail, it was just a facade of HTML..."
— Itamar Gilad
3

Always test multiple design options

Testing a single design is ineffective for predicting behavior. Users could love or hate any design; you need relative comparison. Always present multiple variations to measure which one best drives the target behavior.

Featured guest perspectives
"We never do a UX study where we're just showing people one thing because they could really like it or hate it, but they could really like or hate all the designs. We have no idea. So, when we're doing this, we always present multiple options."
— Kristen Berman
4

Don't skip small, tactical optimizations

Micro-level usability testing can drive millions of dollars in value. Look for 'scary' or confusing CTAs that might be blocking the conversion funnel. Small changes to button text or labels often have outsized impact on business metrics.

Featured guest perspectives
"The micro level, there's so much business value to be derived there... We changed the text on the button with help from our amazing content design... We basically changed seven characters and made Airbnb millions of dollars, because what we found out was really simple. It was just like, 'Hey, this button feels scary.'"
— Judd Antin
5

Engage the whole team in live sessions

Increase team engagement by having cross-functional teams live-react to usability sessions in a shared chat thread. Create a dedicated Slack thread for each live session. Encourage engineers and designers to share real-time observations and pain points.

Featured guest perspectives
"What we wound up doing, especially in the pandemic when we first went remote, is now you can dial into usability sessions and to make it really attractive for the team, what we would do is have people live in a thread, write their real time thoughts... Then you wind up having the PMs, the engineers, designers and the user researcher all in one Slack thread live, responding, reacting to usability session."
— Noah Weiss

Common Mistakes

  • Testing only one design option instead of comparing multiple
  • Relying on dogfooding alone without external user validation
  • Dismissing usability testing as 'junior' work when it's high-leverage
  • Not including PMs and designers in live testing sessions

Signs You're Doing It Well

  • You regularly discover unexpected behaviors that improve product decisions
  • Small optimizations from testing drive measurable business impact
  • The whole team participates in and cares about usability research
  • You can run from idea to prototype to user test within a single day

All Guest Perspectives

Deep dive into what all 11 guests shared about usability testing.

Amjad Masad 1 quote
Listen to episode →
"We've seen product managers build, like I said, like a v1 of an app and actually go out and test it with users. I can't name the company, but there's a public company that have used Replit to test a v1 of an app. And obviously after that sort of works, they take it to the engineers and they're like, 'Okay, we built this thing. We think it's a great thing. We test it with some users.'"
Tactical:
  • Build a functional v1 independently to validate ideas with users before putting them on the official engineering roadmap.
View all skills from Amjad Masad →
Bob Baxley 1 quote
Listen to episode →
"Go observe people going through self-checkout at Target... then go watch it at some other grocery store where it's not as great and really notice what happens with people... Just go watch somebody over 70 fumble with a chip card insert or watch somebody try to figure out Apple Pay."
Tactical:
  • Conduct 'reality checks' by watching users interact with competitor products to see their unbiased behavior and needs.
View all skills from Bob Baxley →
Grant Lee 1 quote
Listen to episode →
"We would have an idea in the morning, come up with some sort of functional prototype, recruit a bunch of people that are legitimately good prospective users, but have zero skin in the game, ship fast so people can start playing with it. In the afternoon, we're already running pretty full scale experiment. You start actually hearing other people describe their usage of the product. We can also watch them struggle. By the evening or by the next day. We can actually go through all of it together and say, okay, we're going back and we have to fix this."
Tactical:
  • Recruit prospective users with 'zero skin in the game' to ensure honest feedback.
  • Watch users struggle and listen to them describe their thought process in real-time.
  • Use platforms like Voicepanel or UserTesting to automate and scale the recruitment process.
View all skills from Grant Lee →
Guillermo Rauch 1 quote
Listen to episode →
"Another aspect of exposure hours is that you tend to overrate how well your products work. It's very important to give your product to another person and watch them interact with it, expose yourself to the pain of reality. And the more you submerge yourself in the real deal, nitty-gritty of what happens when people use your interfaces and whatnot, I think you you'll come out stronger."
Tactical:
  • Invite customers to demo how they use the product live to the executive team or the whole company.
  • Watch for 'pain points' or non-intuitive behaviors that aren't captured in automated metrics.
View all skills from Guillermo Rauch →
Itamar Gilad 2 quotes
Listen to episode →
"Initially you fake it, you do a fake door test, you do a smoke test, Wizard of Oz tests. We used a lot of those in the tabbed inbox by the way, one of the first early versions was actually we showed the tabbed inbox working to people. But it wasn't really Gmail, it was just a facade of HTML and behind the scenes... some of us moved just the subject and the sender into the right place."
Tactical:
  • Run 'Wizard of Oz' tests where humans manually perform the automated task behind a facade
  • Use 'Fake Door' tests to measure user intent and click-through rates on non-existent features
"Initially you fake it, mid-level tests are about building a rough version of it... those are early adopter programs, alphas, longitudinal user studies and fish food. Fish food is testing on your own team."
Tactical:
  • Implement 'Fish Fooding' to catch immediate bugs and UX flaws within the core team
  • Run longitudinal studies to see how user behavior changes over time before a full launch
View all skills from Itamar Gilad →
Judd Antin 2 quotes
Listen to episode →
"The micro level, there's so much business value to be derived there... We changed the text on the button with help from our amazing content design... We basically changed seven characters and made Airbnb millions of dollars, because what we found out was really simple. It was just like, 'Hey, this button feels scary.'"
Tactical:
  • Don't dismiss usability testing as 'junior' work; it is high-leverage for business metrics.
  • Look for 'scary' or confusing CTAs that might be blocking the conversion funnel.
"Doing product walkthroughs to identify lists of potential issues is a great thing to do. Prioritizing that list, figuring out which ones are more or less a problem, and for whom is an area where you should be extremely wary of relying on your own opinion... Some things with a product... you need a pulse to recognize."
Tactical:
  • Use dogfooding to create a list of potential issues, but use research to prioritize them.
  • Acknowledge that your internal intuition is biased by your knowledge of the product.
View all skills from Judd Antin →
Kristen Berman 1 quote
Listen to episode →
"We never do a UX study where we're just showing people one thing because they could really like it or hate it, but they could really like or hate all the designs. We have no idea. So, when we're doing this, we always present multiple options, and then relatively look for which one is going to drive the behavior we're intending to change."
Tactical:
  • Always present multiple design variations in user tests to compare which one best drives the target behavior.
View all skills from Kristen Berman →
Melanie Perkins 1 quote
Listen to episode →
"It's amazing to me how you can find 10 random people on the internet and they can give such astute feedback that then is so representative for such a large number of people."
Tactical:
  • Run tests with as few as 10 random people to identify core product issues
  • Use online platforms like UserTesting.com to get frank, unbiased feedback from users in their own environment
View all skills from Melanie Perkins →
Noah Weiss 1 quote
Listen to episode →
"What we wound up doing, especially in the pandemic when we first went remote, is now you can dial into usability sessions and to make it really attractive for the team, what we would do is have people live in a thread, write their real time thoughts... Then you wind up having the PMs, the engineers, designers and the user researcher all in one Slack thread live, responding, reacting to usability session."
Tactical:
  • Create a dedicated Slack thread for each live usability session.
  • Encourage engineers and designers to share real-time observations and pain points during the session.
View all skills from Noah Weiss →
Upasna Gautam 2 quotes
Listen to episode →
"I had a big working session planned with my users to do research with them, or do user testing, and breaking news breaks, and it takes so much time and effort to gather a team of editors across the globe to do a user testing session. And when breaking news happens, they have to prioritize that over everything."
Tactical:
  • Build in buffers and backups for research sessions involving busy stakeholders
  • Be prepared to pivot or reschedule at a moment's notice
"We create a script and do a simulation of a breaking news scenario to stress test our platform, because all breaking news scenarios are definitely not the same either. This gives us a lot of great feedback in that short amount of time at the speed of breaking news."
Tactical:
  • Script realistic scenarios for users to play out in the product
  • Have engineers and support teams observe the simulation in real-time to identify friction points
View all skills from Upasna Gautam →
Uri Levine 1 quote
Listen to episode →
"Watch new users. Simply watch users and see what they're doing. And number two, if they're not doing what you expect them to do, then ask them why, because this why is the one that is going to make your product successful."
Tactical:
  • Observe users without intervening to see how they actually use the product.
  • Ask 'why' when a user deviates from the expected path.
View all skills from Uri Levine →

Install This Skill

Add this skill to Claude Code, Cursor, or any AI coding assistant that supports Agent Skills.

1

Download the skill

Download SKILL.md
2

Add to your project

Create a folder in your project root and add the skill file:

.claude/skills/usability-testing/SKILL.md
3

Start using it

Claude will automatically detect and use the skill when relevant. You can also invoke it directly:

Help me with usability testing