Noisy rooms often make conversation impossible, but new AI headphones can now automatically mute strangers by learning the unique cadence of the wearer’s discussion.
Researchers at the University of Washington have developed a “proactive listening” system that isolates conversation partners within seconds, solving the “cocktail party problem” without requiring users to look at speakers or fiddle with apps.
The prototype uses off-the-shelf noise-cancelling headphones equipped with artificial intelligence that detects the natural turn-taking rhythm of a conversation. Once the wearer speaks, the system identifies their partners and suppresses all other voices and background noise.
“Our insight is that when we’re conversing with a specific group of people, our speech naturally follows a turn-taking rhythm,” said senior author Shyam Gollakota, a UW professor in the Paul G. Allen School of Computer Science & Engineering. “And we can train AI to predict and track those rhythms using only audio, without the need for implanting electrodes.”
Above the hubbub
Unlike previous hearing assistance technologies that required users to stare at a speaker or manually select a listening zone, this system infers intent automatically. The AI separates the wearer’s group from the surrounding hubbub using a “who spoke when” analysis, effectively creating a clean audio feed of only the relevant participants.
“What we’ve demonstrated is a technology that’s proactive — something that infers human intent noninvasively and automatically,” said lead author Guilin Hu, a doctoral student in the Allen School.
The system can currently handle one to four conversation partners simultaneously without confusing audio lag. In user testing, participants rated the AI-filtered audio more than twice as favourably as the unfiltered baseline.
While dynamic conversations with interruptions or new participants still present challenges, the researchers aim to shrink the technology onto a chip small enough for earbuds or hearing aids.