Siri Activation
People want Siri to be easy to talk to, but unobtrusive. To put it another way, you want it when you want it, and you don’t when you don’t. This poses challenges given that Siri’s hardware button is overloaded — short press for screen lock, double-press for Wallet, and long press for Siri. And using just your voice to summon Siri is natural and convenient, but that convenience comes at the cost of occasional false-activation blunders.
Finding the right balance between these opposing poles — accessibility and unobtrusiveness — is at the heart of my work as the designer responsible for the mechanics of the conversation with Siri: when is the system listening, for how long, and how do I understand when it’s ready for my input — or, hopefully rarely, that it needs a moment to process what I said?
The goal is, of course, an interface that feels anything but mechanical. There’s a surprising amount of technical sophistication behind the simple removal of the requirement to say “Hey” before “Siri” (debuting for English in iOS 17), but the gain in fluidity and confidence that this affords users is tremendous.
But doing more with less speech signal is only the tip of the invocation iceberg. My small team and I collaborate with acoustics, computer vision and location/ranging researchers to build sensor fusion techniques that can model natural engagement and disengagement behaviors across a wide range of devices and scenarios.