AI technology is integrated into hearing aids through features like:
Real-time sound processing: AI chips analyze the listening environment and make instantaneous adjustments to optimize sound quality.
Deep Neural Networks (DNNs): These systems are trained on vast datasets of sounds to differentiate between speech and background noise, allowing the hearing aid to focus on what's important. For example, some DNNs are trained on millions of spoken sentences to improve speech identification.
User-intent sensors: Devices like the Oticon Intent use sensors to monitor head and body movements, as well as the acoustic environment, to predict and adjust to your listening needs.
Google Learning and adaptation: AI-powered devices learn from your daily interactions and preferences, continually improving their ability to adapt to your unique needs over time.
In addition to core sound processing, AI in 2025 hearing aids enables a variety of other features:
Health and wellness tracking: Integrated sensors can track physical activity, assess balance, and even detect falls, sending an alert to emergency contacts.
Enhanced connectivity: New standards like Bluetooth LE Audio and Auracast allow for low-power, high-quality audio streaming from smartphones and public sound systems in places like airports and theaters.
Convenience apps: Companion apps for your smartphone let you fine-tune settings, access health data, and even use features like language translation.
Remote care: Teleaudiology is becoming more prevalent, allowing audiologists to remotely adjust hearing aid settings and provide support.
In this video, we explore how AI is showing up in modern hearing aids—from R&D to real-time sound processing. We’ll walk through current technology, review leading products like Phonak Sphere Infinio and ReSound Vivia, and look ahead at what could be possible in the next few years.
⏱️ Timestamps
00:00 – AI is changing everything—hearing aids included
00:32 – Three ways AI is used in hearing aids
01:30 – #1: AI in the design and training process (Oticon, Starkey, Widex)
02:24 – Even budget options like ELEHEAR use AI in development
03:04 – Most hearing aids today are AI-assisted in design
03:53 – #2: AI in user experience (chat assistants, app control, Edge Mode) 0
4:51 – Example: Signia Assistant + Starkey Edge Mode
05:54 – #3: Real-time AI processing is here (Phonak & ReSound)
06:31 – How real-time AI works: onboard chip, trained to separate speech
07:30 – ReSound trained on 13M+ voice samples
08:15 – Is the AI sound better? It depends on the situation
08:56 – Trade-offs: shorter battery life in AI mode
09:30 – AI is promising but still in early stages
10:17 – Future features: boosting familiar voices, voice control
11:47 – Features like meeting transcription and translation