Overview
As a UX Designer on the Alexa team, I was tasked with identifying features to differentiate the Echo Buds line in a competitive market. I used generative research to move from a broad brief to a specific focus on personal safety. I designed the voice and gesture interactions for two features intended to become core offerings for the product line.
The Problem
I was tasked with enhancing the value proposition of the Echo Buds line via Alexa’s voice and AI capabilities. The goal was to adapt Alexa’s smart features from the home into a mobile, on-the-go environment. I needed to identify how the assistant could move beyond simple music controls to become a proactive companion that makes the product more essential to a user’s daily routine.
Research
Synthesizing general population surveys revealed a recurring theme: customers wanted technology to help them feel safer while out in the world. I used these insights to define two safety-focused features that utilized Alexa's existing sound detection and emergency response capabilities.
Design
Feature A: "Bodyguard"
This feature provided situational awareness through sound detection. I created a priority list of sounds based on user preferences and existing Alexa models. A major challenge was managing processing latency on the earbuds; I had to focus on sounds where a slight delay in notification still provided value to the user, rather than sounds requiring immediate, split-second intervention.
Feature B: "Discreet SOS"
I designed a way to contact help when a phone is out of reach. Because Alexa could not directly dial 911 due to legal constraints, I integrated existing services like Alexa Guard Plus and Emergency Contact Calling to connect users with help.
The design focused on one-handed accessibility and minimizing false alarms. While I originally proposed a new dedicated gesture, technical limitations on the current hardware led me to adapt the design to a gesture already supported by the earbuds, ensuring the feature could be rolled out to existing customers.
Takeaways
-
Designing in Ambiguity: I used research to define a product roadmap where one didn't previously exist.
-
Navigating Constraints: The project was defined by hardware latency, legal restrictions on emergency calling, and engineering limitations.
-
Screenless Interaction: I relied on haptics, voice, and touch gestures to communicate system states, creating an experience that works without a screen.