[Nfbf-l] This device could help the blind see images with their ears.

Alan Dicey adicey at bellsouth.net
Sat Jul 13 02:47:54 UTC 2013


Dear Friends,
the article below appeared online in TIME magazine yesterday, as appended.
With Best Regards,
God Bless,
Alan
Plantation, Florida

Seeing with Sound - The vOICe
http://www.seeingwithsound.com/winvoice.htm

This device could help the blind see images with their ears.
Daredevil time? This technology sounds even cooler.
By Matt Peckham.
Maybe you've heard of synesthesia, the conflation of one sense or body part 
with another - not a word that tumbles from the tongue in casual 
conversation.
It comes from the transliterated Greek words syn, "together" and aisthesis, 
"perception."
People who experience this peculiar-sounding effect might mentally conjure a 
certain smell when hearing a particular sound, or recollect a type of 
texture when seeing a specific color. The celebrated jazz pianist (and 
longtime host of NPR's Piano Jazz) Marian McPartland, for instance, 
identifies musical keys with detailed colors: "The key of D is daffodil 
yellow, B major is maroon, and B flat is blue," she once told jazz critic 
Whitney Balliett.

Now imagine that we could somehow synthesize synesthesia (there's a 
tongue-twister), training our brains to intentionally conflate what's coming 
in through our five senses, say associating certain sounds with specific 
images, effectively learning to "see" with our ears. Think about what that 
could mean to those who can't see (yet have no trouble hearing) if you could 
do it on a practical level.

Researchers with the University of Bath's psychology department in the U.K. 
have a device that they claim may be able to train the brain to do this 
today,teaching people to conjure mental images of what they're hearing in 
their vicinity. While the research just out deals with blindfolded sighted 
people, the implications for the blind and partially-sighted are 
substantial, the impetus being that by exploiting neuroplasticity - the idea 
that the brain can essentially rewire itself to compensate, whether 
functioning with a missing limb or an absent sensory input - you might be 
able to train a brain to visualize its surroundings just by listening to 
them.

Don't confuse that with echolocation, meaning the ability to locate objects 
using echoes, say tapping a cane to orient yourself in a room, which some 
people can already do today. If you read comics you may be familiar with a 
wildly exaggerated version of echolocation: Stan Lee and Bill Everett's 
Daredevil (Matt Murdock), blinded by a radioactive substance as a child, 
acquires superhuman sonar-like abilities in the bargain that allow him to 
"see" in ways that dramatically transcend the abilities of normal sighted 
people. This isn't that. This is actually cooler.

Dubbed "vOICe" - I'll take a wild stab that the stylized OIC is having some 
fun with the phrase "oh I see" - this "sensory substitution" device was 
designed to help people compensate for the loss of one sense (usually 
vision) by subbing in another, and - this is important - without having to 
undergo invasive surgery.
vOICe works by capturing live images with a video camera, then converting 
those images to sounds, mapping video to audio by correlating factors like 
luminosity to loudness or height to pitch, scanning each frame (refreshed 
once every second or two) from left to right.

At this point, we're still learning how and to what extent you can push the 
brain's neural plasticity, so to get a sense for this with vOICe, the 
research team - led by psychologist Dr. Michael Proulx with the University 
of Bath's Crossmodal Cognition Laboratory - decided to observe how people 
who could see, but wearing blindfolds, responded to a visual test with the 
device.

Sitting before computer monitors, blindfolded, wearing over-the-ear 
headphones, participants listened as vOICe's camera scanned images of the 
letter "E" displayed in various positions and sizes - something known as the 
Snellen Tumbling E test - then created soundscapes based on shading: dark 
areas produced loud sounds, while light areas (in this case, the "white" 
letter itself) were silent.

In a routine eye exam, your visual accuracy is determined by how far you're 
sitting from this sort of chart while still being able to see the "E" (in 
focus); normal vision is considered 20/20. In the vOICe test, where the best 
possible results would be equivalent to 20/400 vision, participants were 
able to approach that figure consistently, whether they'd had prior training 
with the device or not. Interestingly, study participants who'd had musical 
training performed better than those without.

All of which is promising, says Proulx, citing a recent study that found a 
sight restoration technique involving stem cell implants only yielded 20/800 
visual acuity. "Although this might improve with time and provide the 
literal sensation of sight, the affordable and non-invasive nature of The 
vOICe provides another option," he says.

Not that invasive and non-invasive vision restoration techniques need be 
mutually exclusive: "Sensory substitution devices are not only an 
alternative, but might also be best employed in combination with such 
invasive techniques to train the brain to see again or for the first time," 
adds Proulx.

How might this work in practice, further down the road? Imagine someone 
wearing an unobtrusive head-mounted camera (Google Glass, for instance), 
receiving wireless audio information through tiny earbuds: as they turn to 
look in various directions, the device scans images and correlates those 
with soundscapes, the person's brain then translating those into mental 
images of the objects - braille for your ears, if you will.

And what of those who've lost their vision compared to those born without 
sight?
Common sense suggests someone who's seen a chair (with their eyes), then 
lost their vision, stands a better chance of correlating a soundscape of 
that chair with a representative mental image, than, say, someone who's 
never been able to see at all (we haven't yet devised a mechanism for 
accurately mapping three-dimensional images into the mental circuitry of 
those with no visual frame of reference).

Perhaps if this training involved a third sense, say touch, to help build 
out the audio-visual mental vocabulary, you could increase the resolution of 
the sensory substitution. But it may not matter: the bumps that identify 
words to those who can read braille look nothing like the characters or 
words they represent. People who assimilate information using braille are 
still reading.
And in any case, if you've never seen the world, gaining a general, working, 
real time understanding of what's around you by in essence "transliterating" 
your environment on the fly, gleaning "images" of objects from sounds, would 
be an incredible feat by any measure.

(You can read the full study, titled "How well do you see what you hear? The 
acuity of visual-to-auditory sensory substitution" and published in the 
journal Frontiers in Psychology here.)

http://www.frontiersin.org/Cognitive_Science/10.3389/fpsyg.2013.00330/full

Source URL:
http://techland.time.com/2013/07/09/this-device-could-help-the-blind-see-images-with-their-ears/
_______________________________________________ 





More information about the NFBF-L mailing list