. | . |
Researchers reconstruct spoken words as processed in nonhuman primate brains by Staff Writers Providence RI (SPX) Dec 16, 2019
A team of Brown University researchers has used a brain-computer interface to reconstruct English words from neural signals recorded in the brains of nonhuman primates. The research, published in the journal Nature Communications Biology, could be a step toward developing brain implants that may help people with hearing loss, the researchers say. "What we've done is to record the complex patterns of neural excitation in the secondary auditory cortex associated with primates' hearing specific words," said Arto Nurmikko, a professor in Brown's School of Engineering, a research associate in Brown's Carney Institute for Brain Science and senior author of the study. "We then use that neural data to reconstruct the sound of those words with high fidelity. "The overarching goal is to better understand how sound is processed in the primate brain," Nurmikko added, "which could ultimately lead to new types of neural prosthetics." The brain systems involved in the initial processing of sound are similar in humans and non-human primates. The first level of processing, which happens in what's called the primary auditory cortex, sorts sounds according to attributes like pitch or tone. The signal then moves to the secondary auditory cortex, where it's processed further. When someone is listening to spoken words, for example, this is where the sounds are classified by phonemes - the simplest features that enable us to distinguish one word from another. After that, the information is sent to other parts of the brain for the processing that enables human comprehension of speech. But because that early-stage processing of sound is similar in humans and non-human primates, learning how primates process the words they hear is useful, even though they likely don't understand what those words mean. For the study, two pea-sized implants with 96-channel microelectrode arrays recorded the activity of neurons while rhesus macaques listened to recordings of individual English words and macaque calls. In this case, the macaques heard fairly simple one- or two-syllable words - "tree," "good," "north," "cricket" and "program." The researchers processed the neural recordings using computer algorithms specifically developed to recognize neural patterns associated with particular words. From there, the neural data could be translated back into computer-generated speech. Finally, the team used several metrics to evaluate how closely the reconstructed speech matched the original spoken word that the macaque heard. The research showed the recorded neural data produced high-fidelity reconstructions that were clear to a human listener. The use of multielectrode arrays to record such complex auditory information was a first, the researchers say. "Previously, work had gathered data from the secondary auditory cortex with single electrodes, but as far as we know this is the first multielectrode recording from this part of the brain," Nurmikko said. "Essentially we have nearly 200 microscopic listening posts that can give us the richness and higher resolution of data which is required." One of the goals of the study, for which doctoral student Jihun Lee led the experiments, was to test whether any particular decoding model algorithm performed better than others. The research, in collaboration with Wilson Truccolo, a computational neuroscience expert, showed that recurrent neural networks (RNNs) - a type of machine learning algorithm often used in computerized language translation - produced the highest-fidelity reconstructions. The RNNs substantially outperformed more traditional algorithms that have been shown to be effective in decoding neural data from other parts of the brain. Christopher Heelan, a research associate at Brown and co-lead author of the study, thinks the success of the RNNs comes from their flexibility, which is important in decoding complex auditory information. "More traditional algorithms used for neural decoding make strong assumptions about how the brain encodes information, and that limits the ability of those algorithms to model the neural data," said Heelan, who developed the computational toolkit for the study. "Neural networks make weaker assumptions and have more parameters allowing them to learn complicated relationships between the neural data and the experimental task." Ultimately, the researchers hope, this kind of research could aid in developing neural implants the may aid in restoring peoples' hearing. "The aspirational scenario is that we develop systems that bypass much of the auditory apparatus and go directly into the brain," Nurmikko said. "The same microelectrodes we used to record neural activity in this study may one day be used to deliver small amounts of electrical current in patterns that give people the perception of having heard specific sounds."
How humans learnt to dance; from the Chimpanzee Conga Warwick UK (SPX) Dec 13, 2019 Psychologist observing two chimpanzees in a zoo have discovered that they performed a behaviour hitherto never seen, they coordinated together in a rhythmic social ritual. Two chimpanzees housed in a zoo in the US have sparked the question about how human dance evolved after being observed performing a duo dance-like behaviour, similar to a human conga-line. In the paper 'Coupled whole-body rhythmic entrainment between two chimpanzees' published in the Journal Scientific Reports, researchers ... read more
|
|
The content herein, unless otherwise known to be public domain, are Copyright 1995-2024 - Space Media Network. All websites are published in Australia and are solely subject to Australian law and governed by Fair Use principals for news reporting and research purposes. AFP, UPI and IANS news wire stories are copyright Agence France-Presse, United Press International and Indo-Asia News Service. ESA news reports are copyright European Space Agency. All NASA sourced material is public domain. Additional copyrights may apply in whole or part to other bona fide parties. All articles labeled "by Staff Writers" include reports supplied to Space Media Network by industry news wires, PR agencies, corporate press officers and the like. Such articles are individually curated and edited by Space Media Network staff on the basis of the report's information value to our industry and professional readership. Advertising does not imply endorsement, agreement or approval of any opinions, statements or information provided by Space Media Network on any Web page published or hosted by Space Media Network. General Data Protection Regulation (GDPR) Statement Our advertisers use various cookies and the like to deliver the best ad banner available at one time. All network advertising suppliers have GDPR policies (Legitimate Interest) that conform with EU regulations for data collection. By using our websites you consent to cookie based advertising. If you do not agree with this then you must stop using the websites from May 25, 2018. Privacy Statement. Additional information can be found here at About Us. |