New York (IANS): Researchers have developed an app that could help people speak the language of eyes — literally.
The smartphone app that researchers working with Microsoft have developed an interpret eye gestures in real time, decode these gestures into predicted utterances, and facilitate communication.
Called GazeSpeak, the app would help people with amyotrophic lateral sclerosis (ALS), a condition resulting in individuals gradually losing their strength and the ability to speak, eat or move.
As part of the Enable team at Microsoft Research, the scientists developed GazeSpeak to help people with ALS who can move their eyes but cannot speak.
ALS also causes other motor impairments that affect voluntary muscle movement.
According to the researchers, current eye-tracking input systems for people with ALS or other motor impairments are expensive, not robust under sunlight, and require frequent re-calibration and substantial, relatively immobile setups.
Eye-gaze transfer (e-tran) boards, a low-tech alternative, are challenging to master and offer slow communication rates.