Pixels guide the way for visually impaired
21 Mar 2013
Images transformed into pixels and projected onto a headset can help the visually impaired in everyday tasks, such as navigation, route planning and finding objects, according to a study led by researchers from the Keck School of Medicine of USC.
Researchers hope the pixels, which were developed using a video camera and mathematical algorithm, can provide more information and enhance the vision of patients already fitted with retinal implants.
''Blind people with retinal implants can detect motion and large objects and have improved orientation when walking. In most cases, they can also read large letters,'' said James Weiland, associate professor of ophthalmology at the Keck School and lead author of the paper.
''At the moment, retinal implants are still low-resolution. We believe that our algorithm will enhance retinal implants by providing the user with more information when they are looking for a specific item.''
The findings were published in the 1 March issue of the Journal of Neural Engineering.
Involved in the study were 19 healthy subjects, each of whom undertook training to get used to the pixelated vision.
During the study, they were fitted with a head-mounted display (HMD) and took part in three different experiments - walking an obstacle course, finding objects on an otherwise empty table and searching for a particular target in a cluttered environment.
A video camera was mounted onto the HMD, which collected information in the view of the subject. Mathematical algorithms converted the images into pixels, which were then displayed onto the HMD's screen in front of the subject.
The algorithms used intensity, saturation and edge-information from the camera's images to pick out the five most important, or salient, locations in the image. Blinking dots at the side of the display provided the subjects with additional directional cues if needed.
All three of the experiments were performed with and without cues. When subjects used the directional cues, their head movements, the time to complete the task and the number of errors were all significantly reduced.
The subjects learned to adapt to pixelated vision in all of the tasks, suggesting that image-processing algorithms can be used to provide greater confidence to patients when performing tasks, especially in a new environment.
It is possible that the device could be fitted with voice description so that the subjects are provided with cues, such as ''the red target is to the left.''
''We are currently looking to take this a step further with object recognition, so instead of telling subjects that 'the red object is to the left,' it will tell them that 'the soda can you want is to the left,''' Weiland said.