2024-05-19
A new era of intelligent audio enhancement
In our increasingly noisy world, noise-canceling headphones have become a necessity for many, offering a reprieve from the constant din of modern life. However, traditional noise-canceling technology has a significant limitation – it indiscriminately cancels out all sounds, leaving users unaware of potentially important information in their surroundings.
Enter Shyam Gollakota, an expert in using AI tools for real-time audio processing from the University of Washington. Gollakota and his team have developed a groundbreaking system that allows for targeted speech hearing in noisy environments, revolutionizing the way we interact with audio devices.
Their AI-based headphones are designed to selectively filter out specific sounds while preserving others, granting users an unprecedented level of control over their audio experience. Imagine being able to enjoy the soothing sounds of chirping birds in a park while effortlessly blocking out the chatter of nearby groups – a scenario that was previously unattainable.
"Imagine you are in a park, admiring the sounds of chirping birds, but then you have the loud chatter of a nearby group of people who just can't stop talking," Gollakota explains. "Now imagine if your headphones could grant you the ability to focus on the sounds of the birds while the rest of the noise just goes away. That is exactly what we set out to achieve with our system."
The key to this innovative technology lies in the combination of noise-canceling technology with a smartphone-based neural network trained to identify 20 different environmental sound categories, including alarm clocks, crying babies, sirens, car horns, and birdsong. When a user selects one or more of these categories, the software identifies and plays those sounds through the headphones in real time while filtering out everything else.
Achieving this level of seamless audio enhancement was no easy feat. Gollakota and his team faced the challenge of developing a high-level intelligence capable of identifying all the different sounds in an environment, separating the target sounds from interfering noises, and ensuring that the extracted sounds sync with the user's visual senses in real time.
"To achieve what we want, we first needed a high-level intelligence to identify all the different sounds in an environment," Gollakota explains. "Then, we needed to separate the target sounds from all the interfering noises. If this is not hard enough, whatever sounds we extracted needed to sync with the user's visual senses, since they cannot be hearing someone two seconds too late. This means the neural network algorithms must process sounds in real time in under a hundredth of a second, which is what we achieved."
The team's AI-powered approach has also been employed to focus on human speech, allowing the algorithm to identify a speaker and isolate their voice from ambient noise in real time for clearer conversations.
Gollakota is excited about the potential of this technology, which he believes represents the future of intelligent audio devices. "We have a very unique opportunity to create the future of intelligent hearables that can enhance human hearing capability and augment intelligence to make lives better," he says.
As the world becomes increasingly noisy, innovations like Gollakota's AI-based headphones offer a glimpse into a future where we can regain control over our audio experiences, tailoring them to our specific needs and preferences. With the power of AI and real-time audio processing, the possibilities for enhancing human hearing and creating more immersive audio environments are truly exciting.
Share with friends:
Write and read comments can only authorized users
Last news