We propose a Gaussian Process Regression (GPR) framework to predict perceptual thresholds at unsampled locations while leveraging uncertainty estimates to guide adaptive sampling.
Rethinking sight restoration through models, data, and lived experience.
We are an interdisciplinary group exploring the science of human, animal, and artificial vision. Our passion lies in understanding how vision works, and how it might be restored in people living with incurable blindness. We focus especially on the emerging field of bionic vision, where insights from neuroscience and engineering converge to inform next-generation neurotechnology.
At the heart of our lab is a diverse team that brings together computer science, psychology, and neuroscience. We are united by a fascination with visual perception and a commitment to research that bridges theory and application. Our work spans computational modeling, psychophysics, and machine learning; from studying how individuals with visual impairment perceive the world, to building biophysical models of brain activity, to using virtual and augmented reality as testbeds for new visual accessibility tools.
What sets our lab apart is our close collaboration with implant developers and bionic eye recipients. We aim to unify efforts across the field by creating open-source tools and standardized evaluation methods that can be used across devices and patient populations. Our goal is not only to advance scientific understanding, but to help reshape how vision restoration technologies are conceptualized, tested, and translated—ultimately supporting greater independence and quality of life for people with vision loss.
We propose a Gaussian Process Regression (GPR) framework to predict perceptual thresholds at unsampled locations while leveraging uncertainty estimates to guide adaptive sampling.
Roksana Sadeghi, Michael Beyeler IEEE EMBC ‘25
We evaluate HILO using sighted participants viewing simulated prosthetic vision to assess its ability to optimize stimulation strategies under realistic conditions.
Eirini Schoinas, Adyah Rastogi, Anissa Carter, Jacob Granley, Michael Beyeler IEEE EMBC ‘25
We evaluated whether gamified training improves compensation for population-coding distortions in sight recovery by testing transfer of learning between a dichoptic object recognition task and a filtered version of Fruit Ninja, finding no significant transfer and suggesting limited generalizability of gamification-based rehabilitation.
Rebecca B. Esquenazi, Kimberly Meier, Michael Beyeler, Drake Wright, Geoffrey M. Boynton, Ione Fine Journal of Vision
We present insights from 16 semi-structured interviews with individuals who are either legally or completely blind, highlighting both the current use and potential future applications of technologies for home-based iADLs.
Lily M. Turkstra, Tanya Bhatia, Alexa Van Os, Michael Beyeler Scientific Reports
We propose a perceptual stimulus encoder based on convolutional neural networks that is trained in an end-to-end fashion to predict the electrode activation patterns required to produce a desired visual percept.
Lucas Relic, Bowen Zhang, Yi-Lin Tuan, Michael Beyeler ACM Augmented Humans (AHs) ‘22
We combined deep learning-based scene simplification strategies with a psychophysically validated computational model of the retina to generate realistic predictions of simulated prosthetic vision.
Nicole Han, Sudhanshu Srivastava, Aiwen Xu, Devi Klein, Michael Beyeler ACM Augmented Humans (AHs) ‘21
We propose a Gaussian Process Regression (GPR) framework to predict perceptual thresholds at unsampled locations while leveraging uncertainty estimates to guide adaptive sampling.
Roksana Sadeghi, Michael Beyeler IEEE EMBC ‘25
We evaluate HILO using sighted participants viewing simulated prosthetic vision to assess its ability to optimize stimulation strategies under realistic conditions.
Eirini Schoinas, Adyah Rastogi, Anissa Carter, Jacob Granley, Michael Beyeler IEEE EMBC ‘25
We introduce two computational models designed to accurately predict phosphene fading and persistence under varying stimulus conditions, cross-validated on behavioral data reported by nine users of the Argus II Retinal Prosthesis System.
Yuchen Hou, Laya Pullela, Jiaxin Su, Sriya Aluru, Shivani Sista, Xiankun Lu, Michael Beyeler IEEE EMBC ‘24
(Note: YH and LP contributed equally to this work.)
We present a series of analyses on the shared representations between evoked neural activity in the primary visual cortex of a blind human with an intracortical visual prosthesis, and latent visual representations computed in deep neural networks.
Jacob Granley, Galen Pogoncheff, Alfonso Rodil, Leili Soo, Lily M. Turkstra, Lucas Nadolskis, Arantxa Alfaro Saez, Cristina Soto Sanchez, Eduardo Fernandez Jover, Michael Beyeler Workshop on Representational Alignment (Re-Align), ICLR ‘24
(Note: JG and GP contributed equally to this work.)
We present a systematic literature review of 227 publications from 106 different venues assessing the potential of XR technology to further visual accessibility.
Justin Kasowski, Byron A. Johnson, Ryan Neydavood, Anvitha Akkaraju, Michael Beyeler Journal of Vision 23(5):5, 1–24
(Note: JK and BAJ are co-first authors.)
What is the required stimulus to produce a desired percept? Here we frame this as an end-to-end optimization problem, where a deep neural network encoder is trained to invert a known, fixed forward model that approximates the underlying biological system.
Jacob Granley, Lucas Relic, Michael Beyeler 36th Conference on Neural Information Processing Systems (NeurIPS) ‘22
We optimize electrode arrangement of epiretinal implants to maximize visual subfield coverage.
Ashley Bruce, Michael Beyeler Medical Image Computing and Computer Assisted Intervention (MICCAI) ‘22