Building hearing aids with a Linux-based open hardware board

Open source platform for improving hearing aids uses a BeagleBone Black for its next iteration.
355 readers like this.
Open innovation

Since first published the story of the GNU/Linux hearing aid research platform in 2010, there has been an explosion in the availability of miniature system boards, including the original BeagleBone in 2011 and the Raspberry Pi in 2012. These ARM processor devices built from cellphone chips differ from the embedded system reference boards of the past—not only by being far less expensive and more widely available—but also because they are powerful enough to run familiar GNU/Linux distributions and desktop applications.

What took a laptop to accomplish in 2010 can now be achieved with a pocket-sized board costing a fraction as much. Because a hearing aid does not need a screen and a small ARM board's power consumption is far less than a typical laptop's, field trials can potentially run all day. Additionally, the system's lower weight is easier for the end user to wear.

The openMHA project—from the Carl von Ossietzky Universität Oldenburg in Germany, BatAndCat Sound Labs in Palo Alto, California, and HörTech gGmbH—is an open source platform for improving hearing aids using real-time audio signal processing. For the next iteration of the research platform, openMHA is using the US$ 55 BeagleBone Black board with its 1GHz Cortex A8 CPU.

The BeagleBone family of boards enjoys guaranteed long-term availability, thanks to its open hardware design that can be produced by anyone with the requisite knowledge. For example, BeagleBone hardware variations are available from community members including SeeedStudio and SanCloud.

BeagleBone Black

The BeagleBone Black is open hardware finding its way into research labs.

Spatial filtering techniques, including beamforming and directional microphone arrays, can suppress distracting noise, focusing audio amplification on the point in space where the hearing aid wearer is looking, rather than off to the side where a truck might be thundering past. These neat tricks can use two or three microphones per ear, yet typical sound cards for embedded devices support only one or two input channels in total.

Fortunately, the McASP communication peripheral in Texas Instruments chips offers multiple channels and support for the I2S protocol, originally devised by Philips for short digital audio interconnects inside CD players. This means an add-on "cape" board can hook directly into the BeagleBone's audio system without using USB or other external interfaces. The direct approach helps reduce the signal processing delay into the range where it is undetectable by the hearing aid wearer.

The openMHA project uses an audio cape developed by the Hearing4all project, which combines three stereo codecs to provide up to six input channels. Like the BeagleBone, the Cape4all is open hardware with design files available on GitHub.

The Cape4all, presented recently at the Linux Audio Conference in Berlin, Germany, runs at a sample rate from 24kHz to 96Khz with as few as 12 samples per period, leading to internal latencies in the sub-millisecond range. With hearing enhancement algorithms running, the complete round-trip latency from a microphone to an earpiece has been measured at 3.6 milliseconds (at 48KHz sample rate with 16 samples per period). Using the speed of sound for comparison, this latency is similar to listening to someone just over four feet away without a hearing aid.


The Cape4all might be the first multi-microphone hearing aid on an open hardware platform.

The next step for the openMHA project is to develop a Bluetooth Low Energy module that will enable remote control of the research device from a smartphone and perhaps route phone calls and media playback to the hearing aid. Consumer hearing aids support Bluetooth, so the openMHA research platform must do so, too.

Also, instructions for running a stereo hearing aid on the Raspberry Pi were released by an openMHA user-project.

As evidenced by the openMHA project, open source innovation has transformed digital hearing aid research from an esoteric branch of audiology into an accessible open science.

Daniel James
Daniel James is the director of 64 Studio Ltd, a company developing GNU/Linux products for OEMs and R&D labs. He was one of the founders of the consortium, which promotes the use of GNU/Linux and Free Software in the professional audio field.
User profile image.
Chief Engineer at 64 Studio Ltd. We specialize in Embedded Software for audio & multimedia applications with a focus on Real-time Linux & Debian.


"...Also, instructions for running a stereo hearing aid on the Raspberry Pi were released by an openMHA user-project..."

...and we're supposed to believe that the Raspberry Pi can compete with machines which essentially run machine code in order to achieve the high speeds needed for this type of processing? The RPi is NOT a high-speed machine; it has a high-speed CPU, bogged down by having to run under a high-level language of some sort; it can NOT run true assembly language.

Hi, the I2S audio interface of the BeagleBone or Raspberry Pi can run with very low latency in full-duplex mode, even when using C or C++ plugins. The goal of the OpenMHA project is to evaluate portable code in field trials without premature optimisation. Please bear in mind that research devices like these don't need to be shrunk to the size of (or have the battery life of) commercially available hearing aids for end users.

In reply to by B. R. Bone (not verified)

Creative Commons LicenseThis work is licensed under a Creative Commons Attribution-Share Alike 4.0 International License.