Keywords: Sensory Neuroprostheses, Bionic Vision, Brain Computer Interfaces, BCI, Stimulus Encoding, Autoencoder, Perception, Computational Modeling, Inverse Problem, Retinal Prostheses, Argus, Cortical Prostheses, Vision
TL;DR: We propose an encoder-decoder based stimulus encoding framework for sensory neuroprostheses and demonstrate its effectiveness for visual prostheses.
Abstract: Sensory neuroprostheses are emerging as a promising technology to restore lost sensory function or augment human capabilities. However, sensations elicited by current devices often appear artificial and distorted. Although current models can predict the neural or perceptual response to an electrical stimulus, an optimal stimulation strategy solves the inverse problem: what is the required stimulus to produce a desired response? Here, we frame this as an end-to-end optimization problem, where a deep neural network stimulus encoder is trained to invert a known and fixed forward model that approximates the underlying biological system. As a proof of concept, we demonstrate the effectiveness of this Hybrid Neural Autoencoder (HNA) in visual neuroprostheses. We find that HNA produces high-fidelity patient-specific stimuli representing handwritten digits and segmented images of everyday objects, and significantly outperforms conventional encoding strategies across all simulated patients. Overall this is an important step towards the long-standing challenge of restoring high-quality vision to people living with incurable blindness and may prove a promising solution for a variety of neuroprosthetic technologies.
Supplementary Material: zip