Please note that scientists make speech from brain signals.
Scientists report that they have developed a voice for virtual prostheses – a system that decodes the vocal intentions of the brain and converts them primarily into understandable speech, without the need to move the muscles, even those in the mouth. (Physicist and author Stephen Hawking used his cheek muscles to write keyboard characters that the computer synthesized in his speech.)
"It's a tremendous job, and it moves to another level to restore speech," decoding the brain signals, said Dr. Anthony Ritaccio, Neurologist and Neurologist at Mayo Clinic, Jacksonville, Fla. .
The new system, described on Wednesday in Nature, deciphers brain motor command for vocal movement control during speech – tongue tap, lip narrowing – and creates understandable sentences that bring the speaker's natural cadence closer.
Implant implant systems have given about eight words per minute. The new program generates approximately 150 minutes of natural speech.
Researchers also found that a synthesized voice system based on single-person brain activity could be used and adapted by someone else – an indication that virtual virtual systems may be available within one day.
The team plans to move to clinical trials to further test the system. The biggest clinical challenge can be finding suitable patients: strokes that disable human speech often damage or destroy brain areas that support speech articulation.
That's exciting research. Congratulations to the researchers.
Mike "Mish" Shedlock