Subvocal Speech

In preliminary experiments, NASA scientists found that button-sized sensors,
stuck under the chin and on either side of the "Adam's apple," could gather
nerve signals, transmit them to a processor, and then deliver them to a
computer program that translates them into words. 

"A person using the subvocal system thinks of phrases and talks to himself so
quietly, it cannot be heard, but the tongue and vocal chords do receive
speech signals from the brain," Chuck Jorgensen, a scientist whose team is
developing silent, subvocal speech recognition at NASA's Ames Research Center
explained. 

Eventually, such "subvocal speech" systems could be used in spacesuits, in
noisy places like airport towers to capture air-traffic controller commands,
or even in traditional voice-recognition programs to increase accuracy. A
spin-off would be that handicapped persons could use this system for a
variety of tasks. 

Initial word recognition results were an average of 92% accurate. Scientists
are testing new, "noncontact" sensors that can read muscle signals even
through a layer of clothing. 

Read the complete story at:
http://www.nasa.gov/centers/ames/news/releases/2004/subvocal/subvocal.html

---

NASA Develops 'Mind-Reading' System
From: New Scientist Magazine - 03/18/2004
By: Maggie McKee

NASA scientists have developed a software program that can analyze the nerve
signals in the mouth and throat to read a person's mind. "Biological signals
arise when reading or speaking to oneself with or without actual lip or
facial movement," says Chuck Jorgensen, a neuroengineer at NASA's Ames
Research Center in Moffett Field, CA, who heads the research effort. The NASA
researchers have used the computer program to perform simple searches on the
Web. Button-sized sensors, placed under the chin and on either side of the
Adam's apple, were able to pick up nerve signals from the tongue, throat, and
vocal cords. NASA researchers trained the software program to recognize six
words, and the system was able to pick up the word thought by participants
hooked up to sensors 92 percent of the time. Moreover, they programmed the
system with a matrix of the letters of the alphabet with each column and row
assigned a single-digit number, then had participants silently spell "NASA"
into a Web search engine, which enabled them to browse the Web without using
a keyboard. The system eventually could be used to send commands to rovers on
other planets, or help an injured astronaut to control machines. The
handicapped, including people who are unable to talk, might be able to
benefit from the system as well. University of Sheffield computer scientist
Phil Green says the technology is "interesting and novel," and although he
says it needs more testing, it could be coupled with existing
voice-recognition technology for greater effect. 

Read the entire story at:
http://www.newscientist.com/news/news.jsp?id=ns99994795
