What if it were possible to decode the inner language of individuals deprived of the ability to express themselves? This is the objective of a team of neuroscientists from the University of Geneva (UNIGE) and the University Hospitals of Geneva (HUG), according to an announcement on the University’s website. After more than four years of research, this team has managed to identify promising neural signals to capture our internal monologues. It was also able to identify the brain areas to be observed in priority to try to decipher them in the future. These results open new perspectives for the development of interfaces for people suffering from aphasia. They can be found in the journal Nature Communications.
When human beings speak, different areas of their brain must be activated. However, the function of these regions can be seriously impaired after damage to the nervous system. For example, amyotrophic lateral sclerosis (or Charcot’s disease) can completely paralyze the muscles used to speak. In other cases, following a stroke for example, areas of the brain responsible for language can be affected: this is called aphasia. However, in many of those cases, the ability of patients to imagine words and sentences remains partly functional.
Decoding our internal speech is therefore of great interest to neuroscience researchers. But the task is far from easy, as Timothée Proix, scientist in the Department of Basic Neuroscience at the UNIGE Faculty of Medicine, explained: “Several studies have been conducted on the decoding of spoken language, but much less on the decoding of imagined speech. This is because, in the latter case, the associated neural signals are weak and variable compared to explicit speech. They are therefore difficult to decode by learning algorithms.” That is, through computer programs.
When a person speaks aloud, he or she produces sounds that are emitted at certain precise moments. Researchers can thus relate these tangible elements to the brain regions involved. In the case of imagined speech, the process is much less easy. Scientists have no obvious information on the sequencing and tempo of the words or sentences formulated internally by the individual. The areas recruited in the brain are also less numerous and less active.
In order to perceive the neural signals of this very particular type of speech, the UNIGE team used a panel of 13 hospitalized patients, in collaboration with two American hospitals. They collected data through electrodes implanted directly into patients’ brains in order to assess their epileptic disorders.
“We asked these people to say words and then to imagine them. Each time, we reviewed several frequency bands of brain activity known to be involved in language,” said Anne-Lise Giraud, a professor in the Department of Basic Neuroscience at the UNIGE Faculty of Medicine, newly appointed director of the Institut de l’Audition in Paris, and co-director of the NCCR Evolving Language.
The researchers observed several types of frequencies produced by different brain areas when these patients spoke, either orally or internally.
“First of all, the oscillations called theta (4-8Hz), which correspond to the average rhythm of syllable elocution. Then the gamma frequencies (25-35Hz), observed in the areas of the brain where speech sounds are formed. Thirdly, beta waves (12-18Hz) related to the cognitively more efficient regions solicited, for example to anticipate and predict the evolution of a conversation. Finally, the high frequencies (80-150Hz) that are observed when a person speaks out,” said Pierre Mégevand, assistant professor in the Department of Clinical Neurosciences at the Faculty of Medicine of the UNIGE and associate physician at the HUG.
Thanks to these observations, the scientists were able to show that the low frequencies and the coupling between certain frequencies (beta and gamma in particular) contain essential information for the decoding of imagined speech. Their research also reveals that the temporal cortex is an important area for the eventual decoding of internal speech. Located in the left lateral part of the brain, this specific cerebral region is involved in the processing of information related to hearing and memory, but it also houses a part of Wernicke’s area, responsible for the perception of words and language symbols.
These results are a major advance in the reconstruction of speech from neural activity.
“But we are still a long way from being able to decode imagined language,” concluded the research team.
Original Paper: Proix T, Saa JD, Christen A, et al. Imagined speech can be decoded from low- and cross-frequency intracranial EEG features. Nature Communications. 2022;13(48).
Source: University of Geneva, Nature Communications