[ Back to EurekAlert! ] Public release date: 12-Apr-2007
[ | E-mail Share Share ]

Contact: Rolf Eckmiller
eckmiller@nero.uni-bonn.de
49-228-734-422
University of Bonn

Learning visual prosthesis at the Hanover Fair

Bonn computer scientists demonstrate a system, which improves the function of retinal implants significantly

This release is also available in German.

When the idea appeared several years ago, it sounded persuasive: How about implanting electrodes at the defective retina of blind subjects and connecting them with a mini camera in order to re-establish vision.

However, the first clinical tests yielded quite sobering results: Following the surgery, the patients could not even distinguish simple shapes. Neural computation scientists at Bonn University will introduce a software system at the Hanover Fair that is hoped to change this: with the aid of this software, the visual prosthesis "learns" to generate exactly those signals, which are expected and can be interpreted by the brain. The learning visual prosthesis will be exhibited between April 16th and April 20th at the community booth of the Science Region Bonn (hall 2, booth D35).

Nearly two dozens of patients in Germany and the U.S. have so far been implanted with a visual prosthesis. For this purpose, clinicians open the eye ball and attach a thin foil at the retina. Small protruding contacts reach neurons, which form the ganglion cell layer of the retina. These electrical stimulation contacts feed camera signals into the optic nerve. The camera may be attached to a frame of glasses and transmits its signals in a wireless fashion to the implant.

Currently, the results do not meet the high expectations. "The camera generates electrical signals, which are almost useless for the brain," comments Rolf Eckmiller, a professor at the Department of Computer Science at Bonn University. "Our own system translates the camera signals into a language, which the central visual system in the brain understands". Unfortunately, the central visual system of each individual speaks a different dialect; this poses a difficulty for the translator function. For this reason, the computer- and neural scientist developed the "Retina Encoder" together with his graduate students Oliver Baruth and Rolf Schatten. At the Hanover Fair he is looking for commercial partners for the next step into clinical trials.

Translator in the Spectacles

"In principle, the Retina Encoder is a computer program that converts the camera signals and forwards them to the retinal implant," explains Oliver Baruth the function. "The encoder learns in a continuous process how to change the camera output signal so that the respective patient can perceive the image." Currently, tests of the learning dialog process are being performed with normally sighted volunteers. The camera images are translated by the Retina Encoder and subsequently forwarded to a kind of "virtual central visual system." This simulation mimics the brain function for the interpretation of the converted camera data.

Initially, the Retina Encoder does not know which language the virtual central visual system speaks. Therefore the software translates the original picture - for example a ring - in different, randomly selected "dialects". This way, variations of the picture emerge, which are more or less similar to a ring. The volunteer sees these variations on a small screen that is integrated in a frame of glasses. By means of head movements, the person selects those variations that appear most similar to a ring. From these choices, the learning software draws conclusions how to improve the translation. In the next learning cycle, several new picture variations are being presented, which look already more similar to the original: during this process, the Retina Encoder becomes adapted step-by-step to the language of the virtual central visual system. In the current tests it works very well; however, the scientists have not yet tested their system in patients. The scientists emphasize that in principle, the Retina Encoder could be integrated in implanted visual prostheses within a few months.

In normally sighted humans, a kind of natural Retina Encoder is already integrated in the retina: specifically, four layers of nerve cells are positioned in front of the photoreceptor cells. "The retina is a transparent biocomputer," Eckmiller says. "It transforms the electrical signals of rod and cone photoreceptors into a complex signal." This signal reaches the brain via the optic nerve.

Don't expect Miracles

In the brain, the complex information is being decoded. The brain acquires the corresponding ability within the first months of life. During this time, the central visual system becomes individually adjusted to the retinal signals: the brain learns how to interpret the data from the optic nerve. In adults, however, who become blind later in life, the central visual system is already matured: it is not able anymore to change easily. "If the central visual system is not as flexible anymore, the artificial retina has to be," Eckmiller points out. “The artificial retina must learn to generate signals that are useful for the brain. And exactly this learning ability distinguishes our Retina Encoder”.

Nevertheless, he warns against too high expectations," Nobody should think that one could read again ones favorite detective stories with a visual prosthesis. One may perhaps be able to recognize the 'Gestalt' of larger objects and to perceive contours; a higher vision quality can not be expected in the near future. However, for a blind subject this represents a major improvement. He will be able to orient himself again in his environment. This gain in independence is our goal!"

###


[ Back to EurekAlert! ] [ | E-mail Share Share ]

 


AAAS and EurekAlert! are not responsible for the accuracy of news releases posted to EurekAlert! by contributing institutions or for the use of any information through the EurekAlert! system.