[Nfbf-l] Robot Sensing and Smartphones Help Blind Navigate

Alan Dicey adicey at bellsouth.net
Mon Mar 25 00:25:43 UTC 2013


Robot Sensing and Smartphones Help Blind Navigate


ROBOTS need help navigating their surroundings and sophisticated location 
systems to keep track of their position. Now the same technologies are being 
adapted to help blind people navigate indoor and outdoor spaces 
independently.


One such system, being developed by Edwige Pissaloux and colleagues at the 
Institute of Intelligent Systems and Robotics at the Pierre and Marie Curie

University in Paris, France, consists of a pair of glasses equipped with 
cameras and sensors like those used in robot exploration. The system, 
unveiled at a talk at the Massachusetts Institute of Technology this month, 
produces a 3D map of the wearer's environment and their position within it 
that is constantly

updated and displayed in a simplified form on a handheld electronic Braille 
device. It could eventually allow blind people to make their way, unaided, 
wherever they want to go, says Pissaloux. "Navigation for me means not only 
being able to move around by avoiding nearby obstacles, but also to 
understand how the space is socially organised - for example, where you are 
in relation to the pharmacy, library or intersection," she says.


Two cameras on either side of the glasses generate a 3D image of the scene. 
A processor analyses the image, picking out the edges of walls or objects, 
which it uses to create a 3D map. The system's collection of accelerometers 
and gyroscopes - like those used in robots to monitor their position - keeps 
track of the user's location and speed. This information is combined with 
the image to determine the user's position in relation to other objects.


The system generates almost 10 maps per second which are transmitted to the 
handheld Braille device to be displayed as a dynamic tactile map. The 
Braille pad consists of an 8-centimetre-square grid of 64 taxels - pins with 
a shape memory alloy spring in the middle. When heat is applied to the 
springs, they expand, raising the pins to represent boundaries. The Braille 
version of the map is updated fast enough for a visually-impaired wearer to 
pass through an area at walking speed, says Pissaloux. Seth Teller, who 
develops assistive technologies at MIT, calls the work exciting and 
ambitious.


This is not the only robotics project to be re-purposed. Software that 
predicts how far a robot has travelled based on information from its 
on-board sensors is being modified to track a person's movements based on 
their stride length. The low-cost system, being developed by Eelke Folmer 
and Kostas Bekris at the University of Nevada in Reno would help blind 
people navigate around buildings using just a smartphone.


The new system uses freely available 2D digital indoor maps and the 
smartphone's built-in accelerometer and compass. Directions are provided 
using synthetic speech. To help the smartphone calibrate and adjust to a 
user's individual stride length, the user must initially use touch to detect 
the landmarks in their environment, such as corridor intersections, doors 
and elevators. The system will be presented at the IEEE International 
Conference on Robotics and Automation in St Paul, Minnesota, in May.


David Ross at the Atlanta Vision Loss Center in Decatur, Georgia, says that 
the sensing problems faced by robots and blind people are similar but there 
are big differences. "Sensing systems developed for mobile robots may have 
some application, but must be adapted considerably to suit a wide variety of 
human needs and situations," he says.


The all-seeing ring

A virtual assistant can help blind people explore their surroundings. 
Developed by Suranga Nanayakkara at the MIT's Media Lab, EyeRing consists of 
a ring equipped with a camera, and headphones. The user points the ring at 
an object they are holding and uses voice commands to say what they need to 
know - the colour of an item of clothing, say, or the denomination of paper 
money. The ring takes a picture of the object, which is transmitted 
wirelessly to

a cellphone, where software analyses the image. The required information is 
then read out by a synthesised voice. It is being presented at the 
Conference

on Human Factors in Computing Systems in Austin, Texas.


__._,_.___ 





More information about the NFBF-L mailing list