Overview
APPLE SIRI

Apple Siri’s speech recognition has presented significant racial and gender bias. This is a redesign of Siri's settings and feature to improve speech recognition accuracy and performance. Our team's goal was to re-program Siri to better understand the diverse voices of each of their users.

Duration

10 Weeks (Jan - Mar 2021)

Role

UX Designer and Researcher

Team

Zerabruk Abraham
Beneyam Mesele
Natalie Roy

the problem

Siri misinterprets words spoken by users whose first language is not English.

Apple Siri uses speech recognition to allow users to perform tasks on their iPhone, such as speech-to-text for texting, playing music, or answering questions from the web. Their current dataset does not recognize speech spoken by users whose first language is not English or users who don’t identify as a white American. They are forced to eeh-nuhn-see-eyt each one of her words in order for Siri to understand.

Claudia lopez lloreda
Siri user from puerto rico
“Clow-dia,” I say once. Twice.
A third time.
Defeated, I say the Americanized version of my name:
“Claw-dee-ah.”
Finally, Siri recognizes it.
audit
Evaluating the Current Experience

One of the main pain points we discovered was that Siri was limited to understanding one language at a time. The user is prompted to select one language and one voice for Siri to respond in. If a user speaks regularly in two languages or has a strong accent in one, there is no option to help Siri better recognize their voice.

Reviewing Existing Research

To narrow the scope of the project, our team cross referenced existing research. This revealed that Siri has a 92% accuracy rate when it comes to understanding white American males, a reflection of their first dataset being military personnel. With over 60% of military personnel being white American males, Apple unknowingly developed their program to have a large disparity in voices recognized.

Siri Speech Recognition Accuracy Rates
User Research
User Testing

Our team wanted to understand the experience of Siri users and their experience with speech recognition. I conducted 10 user interviews with 5 individuals who spoke at least 2 different languages, and 5 individuals who spoke only English and grew up in America. We made sure each of them had an iPhone with Siri installed and asked them to complete 5 tasks on their device with Siri.

5 User Tasks
1.
Schedule a meeting for next friday at 10am
2.
Text angelina "Hi, how are you doing?"
3.
Find the temperature outside in farenheit
4.
increase and decrease the volume
5.
increase and decrease the brightness of screen
Key Takeaways
Accidental triggers of "Hey Siri" are common in diverse individuals.
There is a higher speech recognition accuracy rate for white individuals.
Non-english native speakers felt uncomfortable at times while using Siri.
Design
Low-Fi Ideation

Our first idea was to re-design the settings of Siri by implementing a feature titled “Accent Assumption”. This would allow for people to choose from a longer and more specific list of accents and languages.

Interface aside, our second idea was to include more diverse datasets and training. We proposed that Siri increase their datasets and training programs to include more racial minorities, ethnic identities, and accents.

These changes are outlined in the following prototype.

Solution
Prototype
Learnings
Understand your user

Understand who the target user and demographic is before designing. This will help the team recognize what racial bias, gender bias, or disabilities they should account for.

Don't skip user research

Conduct user research to identify what the needs of the users are.

Collaboration Multiplies Success

Collaborate with everyone. Ideation processes with engineers, management, QAs, and etc. can give new insight and diverse backgrounds.

© Designed by Angelina Lum