With no more than 70 hours of training on average, study participants could recognize the presence of a human form. What's more, they were able to detect the exact posture of the person in the image and imitate it.
"The idea is to replace information from a missing sense by using input from a different sense," explains Amir Amedi of The Hebrew University of Jerusalem. "It's just like bats and dolphins use sounds and echolocation to 'see' using their ears."
"Imagine for instance a diagonal line going down from left to right; if we use a descending musical scalegoing on the piano from right to leftit will describe it nicely," continues Ella Striem-Amit, also of The Hebrew University. "And if the diagonal line is going up from left to right, then we use an ascending musical scale."
The researchers first taught people to perceive simple dots and lines. Then those individuals learned to connect the lines with junctions or curves, gradually working up to more and more complex images.
The success of the sensory substitution approach suggests great potential for its use in therapy more broadly, the researchers say.
"We're beginning to understand [that] the brain is more than a pure sensory machine," Amedi says. "It is a highly flexible task machine. The time has come to revive the focus on practical visual rehabilitation with sensory substitution devices."
In the current study, the researchers used an algorithm they call vOICe, which translates images in black and white. But they've since developed a newer algorithm, EyeMusic, which incorporates color information through the use of different musical instruments. The EyeMusic app is freely available on iTunes, where the researchers hope it can be useful to the blind and the general public alike.