[Nfbf-l] Robot Sensing and Smartphones Help Blind Navigate
Alan Dicey
adicey at bellsouth.net
Mon Mar 25 00:25:43 UTC 2013
Robot Sensing and Smartphones Help Blind Navigate
ROBOTS need help navigating their surroundings and sophisticated location
systems to keep track of their position. Now the same technologies are being
adapted to help blind people navigate indoor and outdoor spaces
independently.
One such system, being developed by Edwige Pissaloux and colleagues at the
Institute of Intelligent Systems and Robotics at the Pierre and Marie Curie
University in Paris, France, consists of a pair of glasses equipped with
cameras and sensors like those used in robot exploration. The system,
unveiled at a talk at the Massachusetts Institute of Technology this month,
produces a 3D map of the wearer's environment and their position within it
that is constantly
updated and displayed in a simplified form on a handheld electronic Braille
device. It could eventually allow blind people to make their way, unaided,
wherever they want to go, says Pissaloux. "Navigation for me means not only
being able to move around by avoiding nearby obstacles, but also to
understand how the space is socially organised - for example, where you are
in relation to the pharmacy, library or intersection," she says.
Two cameras on either side of the glasses generate a 3D image of the scene.
A processor analyses the image, picking out the edges of walls or objects,
which it uses to create a 3D map. The system's collection of accelerometers
and gyroscopes - like those used in robots to monitor their position - keeps
track of the user's location and speed. This information is combined with
the image to determine the user's position in relation to other objects.
The system generates almost 10 maps per second which are transmitted to the
handheld Braille device to be displayed as a dynamic tactile map. The
Braille pad consists of an 8-centimetre-square grid of 64 taxels - pins with
a shape memory alloy spring in the middle. When heat is applied to the
springs, they expand, raising the pins to represent boundaries. The Braille
version of the map is updated fast enough for a visually-impaired wearer to
pass through an area at walking speed, says Pissaloux. Seth Teller, who
develops assistive technologies at MIT, calls the work exciting and
ambitious.
This is not the only robotics project to be re-purposed. Software that
predicts how far a robot has travelled based on information from its
on-board sensors is being modified to track a person's movements based on
their stride length. The low-cost system, being developed by Eelke Folmer
and Kostas Bekris at the University of Nevada in Reno would help blind
people navigate around buildings using just a smartphone.
The new system uses freely available 2D digital indoor maps and the
smartphone's built-in accelerometer and compass. Directions are provided
using synthetic speech. To help the smartphone calibrate and adjust to a
user's individual stride length, the user must initially use touch to detect
the landmarks in their environment, such as corridor intersections, doors
and elevators. The system will be presented at the IEEE International
Conference on Robotics and Automation in St Paul, Minnesota, in May.
David Ross at the Atlanta Vision Loss Center in Decatur, Georgia, says that
the sensing problems faced by robots and blind people are similar but there
are big differences. "Sensing systems developed for mobile robots may have
some application, but must be adapted considerably to suit a wide variety of
human needs and situations," he says.
The all-seeing ring
A virtual assistant can help blind people explore their surroundings.
Developed by Suranga Nanayakkara at the MIT's Media Lab, EyeRing consists of
a ring equipped with a camera, and headphones. The user points the ring at
an object they are holding and uses voice commands to say what they need to
know - the colour of an item of clothing, say, or the denomination of paper
money. The ring takes a picture of the object, which is transmitted
wirelessly to
a cellphone, where software analyses the image. The required information is
then read out by a synthesised voice. It is being presented at the
Conference
on Human Factors in Computing Systems in Austin, Texas.
__._,_.___
More information about the NFBF-L
mailing list