Artificial intelligence is turning brain signals into speech

Artificial intelligence is turning brain signals into speech

Scientists can now read your mind: sort of.

Researchers from the University of California claim to have created a device that turns brain signals into electronic speech. The “brain machine interface” is a neural decoder that maps cortical activity into movements of the vocal tract, with one participant of the study being asked to read sentences aloud, before miming the same sentences with their mouth without producing sound. The researchers described the results as “encouraging”.

Five volunteers had electrodes implanted on the surface of their brains as part of epilepsy treatment. The researchers recorded the brain activity of the study participants reading sentences, before combining these recordings with data on sound creation and how the tongue, lips, jaw and larynx help us speak.

What is Deep Learning?

A deep-learning algorithm was used on the data collected on vocal tract movement and the research from participants’ speech. A deep learning algorithm is a dense artificial neural network, which is in turn inspired by the way that a biological brain works.

Deep learning encompasses machine learning, where machines learn by experience and develop skills without needing the input of their human masters. From colourising images to facial recognition, deep learning is already contextualising the world around it. Converting brain activity into speech is a huge breakthrough, though ordinary speech translations of one language to another are perhaps the best examples of the technology at work.


Deep learning is already contextualising the world around it. Converting brain activity into speech is a huge breakthrough


Deep learning algorithms are actually rethinking the way that we translate languages. Previously, tools such as Google Translate would have translated every word in a sentence individually, as if looking up each word in a dictionary. These days, however, algorithms invented just two years ago can perform to the level of statistical machine translation systems invented 20 years ago, because they can decode sentences using recurrent neural networks (RNN).

RNNs learn patterns in data and output them as encoded sentences. There’s no need to input the rules about human languages because the RNN learns everything it needs to know. This is how deep-learning can solve sequence-to-sequence challenges such as speech translation, whether that’s from one language to another, or from brainwaves into sounds.

How impressive is this study?

Similarly to how online translators began by simply turning individual words from one language to another, scientists have only been able to use AI to analyse brain activity and translate a syllable at a time.

Many who have lost the ability to speak have used speech-generating devices and software. Naturally spoken speech averages 150 words per minute – up to fifteen times more than devices used by motor-neuron disease sufferers – so giving the participants of this study a more effortless flow of speech is a breakthrough.  

“Technology that translates neural activity into speech would be transformative for people who are unable to communicate as a result of neurological impairments,” said the neuroscientists in the study, published on April 24th in Nature. “Decoding speech from neural activity is challenging because speaking requires very precise and rapid multi-dimensional control of vocal tract articulators.”

An electronic mesh, which consists of a network of flexible circuits placed into the brain, is now being tested on animals. Elon Musk’s Neuralink company is also developing an interface between computers and the biological brains, using neural lace technology, in what the company describes as, “ultra high bandwidth brain-machine interfaces to connect humans and computers”.

Artificial Intelligence is working on other senses

Vision is another sense that will benefit in the future from reading neural output.

A recent study has looked at how machine learning can visualise perceptual content, by analysing human functional magnetic resonance imaging (fMRI). The feature decoding analysis was made with fMRI activity patterns in visual cortex (VC) measured while subjects so much as imagined visual images. Decoded features were then sent to the reconstruction algorithm to generate an image.

Similarly to the speech research, study into deep image reconstruction from human brain activity is suggests that artificial neural networks can provide a new window into the internal contents of the brain.

It’s perhaps getting ahead of ourselves to suggest that mind-reading is imminent, but it’s a certainty that artificial intelligence and deep learning will provide the human race with a neuromechanical biological system of sorts. “These findings advance the clinical viability of using speech neuroprosthetic technology to restore spoken communication,” the study said.

Photo from https://www.nature.com/

Luke Conrad

Technology & Marketing Enthusiast

TPIs are the Future of Energy Solutions

David Sheldrake SVP POWWR • 19th June 2025

The energy industry is undergoing a transformation, and Third-Party Intermediaries (TPIs), those brokers and consultants who help businesses procure energy, are at the centre of it. With growing complexity, increasing regulation, and evolving customer expectations, the role of TPIs is shifting from price-focused brokers to strategic energy advisors. While renewable energy adoption continues to reshape...

Quick Commerce and the Retail Media Revolution

Sue Azari • 11th June 2025

Quick commerce has transformed the way consumers shop, redefining convenience with near-instant delivery of groceries, meals, and household essentials. However, beyond its impact on logistics and e-commerce, quick commerce is now emerging as a major force in digital advertising. As consumer behaviours shift toward on-demand purchases, these platforms are leveraging their vast first-party data and...

Is It Time for a VMware Alternative?

Wind River • 22nd May 2025

Companies have options when it comes to replacing VMware as their cloud platform, to address rising costs, support concerns, and a shrinking partner ecosystem. If you are ready to contemplate a different vendor, here are five reasons why Wind River Cloud Platform should be on your short list of VMware alternatives.

AI Leads as VivaTech Unveils Top 100 Startups

Viva Technology • 14th May 2025

Viva Technology has unveiled the first edition of its “Top 100 Rising European Startups for 2025,” spotlighting the most promising young companies shaping Europe’s tech future. Germany, France, and the UK lead the ranking, which highlights high-growth startups across 13 countries. Artificial intelligence dominates the list, with 15 companies spanning AI agents, models, and infrastructure....

Birmingham Unveils the UK’s Best Emerging HealthTech Advances

Kosta Mavroulakis • 03rd April 2025

The National HealthTech Series hosted its latest event in Birmingham this month, showcasing innovative startups driving advanced health technology, including AI-assisted diagnostics, wearable devices and revolutionary educational tools for healthcare professionals. Health stakeholders drawn from the NHS, universities, industry and front-line patient care met with new and emerging businesses to define the future trajectory of...

Why DEIB is Imperative to Tech’s Future

Hadas Almog from AppsFlyer • 17th March 2025

We’ve been seeing Diversity, Equity, Inclusion, and Belonging (DEIB) initiatives being cut time and time again throughout the tech industry. DEIB dedicated roles have been eliminated, employee resource groups have lost funding, and initiatives once considered crucial have been deprioritised in favour of “more immediate business needs.” The justification for these cuts is often the...