Why Neuralink’s Blindsight and brain implants to restore sight won’t work like human eyesight

Elon Musk recently pronounced that the next Neuralink project will be a “Blindsight” cortical implant to restore vision: “Resolution will be low at first, like early Nintendo graphics, but ultimately may exceed normal human vision.”

Unfortunately, this claim rests on the fallacy that neurons in the brain are like pixels on a screen. It’s not surprising that engineers often assume that “more pixels equals better vision.” After all, that is how monitors and phone screens work.

In our newly published research, we created a computational model of human vision to simulate what sort of vision an extremely high-resolution cortical implant might provide. A movie of a cat with a resolution of 45,000 pixels is sharp and clear. A movie generated using a simplified version of a model of 45,000 cortical electrodes, each of which stimulates a single neuron, still has a recognizable cat but most of the details of the scene are lost.

The reason the movie generated by electrodes is so blurry is because neurons in the human visual cortex do not represent tiny dots or pixels. Instead, each neuron has a particular receptive field, which is the location and pattern a visual stimulus must have in order to make that neuron fire. Electrically stimulating a single neuron produces a blob whose appearance is determined by that neuron’s receptive field. The tiniest electrode—one that stimulates a single neuron—will produce a blob that is roughly the size of your pinkie’s width held at arm’s length.

Consider what happens when you look at a single star in the night sky. Each point in space is represented by many thousands of neurons with overlapping receptive fields. A tiny spot of light, such as a star, results in a complex pattern of firing across all these neurons.

To generate the visual experience of seeing a single star with cortical stimulation, you would need to reproduce a pattern of neural responses that is similar to the pattern that would be produced by natural vision.

In order to do this, you would obviously need thousands of electrodes. But you would also need to replicate the correct pattern of neuronal responses, which requires knowing every neuron’s receptive field. Our simulations show that knowing the location of each neuron’s receptive field in space is not enough – if you don’t also know the orientation and size of each receptive field, then the star becomes a fuzzy mess.

So, even a single star—a single, bright pixel—generates an immensely complex neural response in the visual cortex. Imagine the even more complex pattern of cortical stimulation needed to accurately reproduce natural vision.

Some scientists have suggested that by stimulating exactly the right combination of electrodes, it would be possible to produce natural vision. Unfortunately, no one has yet suggested a sensible way to determine the receptive field of each individual neuron in a specific blind patient. Without that information, there is no way to see the stars. Vision from cortical implants will remain grainy and imperfect, regardless of the number of electrodes.

Sight restoration is not simply an engineering problem. Predicting what kind of vision a device will provide requires knowing how the technology interfaces with the complexities of the human brain.

How we created our virtual patients

In our work as computational neuroscientists, we develop simulations that predict the perceptual experience of patients seeking to restore their sight.

We have previously created a model to predict the perceptual experience of patients with a retinal implant. To create a virtual patient to predict what cortical implant patients would see, we simulated the neurophysiological architecture of the area of the brain involved in the first stage of visual processing. Our model approximates how receptive fields increase in size from central to peripheral vision and the fact that every neuron has a unique receptive field.

Our model successfully predicted data describing the perceptual experience of participants across a wide range of studies on cortical stimulation in people. Having confirmed that our model could predict existing data, we used it to make predictions about the quality of vision that possible future cortical implants might produce.

Models like ours are an example of virtual prototyping, which involves using computer systems to improve product design. These models can facilitate new technology development and evaluate device performance. Our study shows they can also offer more realistic expectations about what kind of vision bionic eyes might provide.

First do no harm

In our almost 20 years researching bionic eyes, we’ve seen the complexity of the human brain defeat company after company. Patients pay the cost when these devices fail, left stranded with orphaned technologies in their eye or brain.

The Food and Drug Administration could mandate that sight recovery tech companies must develop failure plans that minimize harm to patients when technologies stop working. Possibilities include requiring companies implanting neuroelectronic devices into patients to participate in technology escrow agreements and carry insurance to ensure continuing medical care and technology support if they go bankrupt.

If cortical implants can achieve anything close to the resolution of our simulations, that would still be an accomplishment worth celebrating. Grainy and imperfect vision would be life-changing for many thousands of people who currently suffer from incurable blindness. But this is a moment for cautious rather than blind optimism.

Ione Fine is a professor of psychology at the University of Washington.

Geoffrey Boynton is a professor of psychology at the University of Washington.

This article is republished from The Conversation under a Creative Commons license. Read the original article.

https://www.fastcompany.com/91168537/neuralink-blindsight-brain-implants-human-eyesight?partner=rss&utm_source=rss&utm_medium=feed&utm_campaign=rss+fastcompany&utm_content=rss

Létrehozva 9mo | 2024. aug. 12. 10:50:03


Jelentkezéshez jelentkezzen be

EGYÉB POSTS Ebben a csoportban

Anthropic hires a top Biden official to lead its new AI-for-social-good team (exclusive)

Anthropic is turning to a Biden administration alum to run its new Beneficial Deployments team, which is tasked with helping extend the benefits of its AI to organizations focused on social good—p

2025. máj. 5. 21:20:03 | Fast company - tech
Speed-limiting devices could be coming for reckless U.S. drivers in these states

A teenager who admitted being “addicted to speed” behind the wheel had totaled two other cars in the year before he slammed into a minivan at 112 mph (180 kph) in a Seattle suburb,

2025. máj. 5. 16:40:03 | Fast company - tech
Nvidia chips could face new tracking rules under a bipartisan bill to stop chip smuggling to China

A U.S. lawmaker plans to introduce legislation in coming weeks to verify the location of

2025. máj. 5. 16:40:02 | Fast company - tech
Meta’s AI social feed is a privacy disaster waiting to happen

Since ChatGPT sparked the generative AI revolution in November 2022, interacting with AI has felt like using a digital confession booth—private, intimate, and shielded from public view (unless you

2025. máj. 5. 14:20:05 | Fast company - tech
I have trouble focusing, but this AI browser feature helps

My worst workday habit is that I’m a compulsive web page checker.

Throughout the day, I’m constantly refreshing the same handful of sites for updates. I’ll check the me

2025. máj. 5. 11:50:07 | Fast company - tech
This is the future of AI, according to Nvidia

​​Recent breakthroughs in generative AI have centered largely on language and imagery—from chatbots that compose sonnets and analyze text to voice models that mimic human speech and tools that tra

2025. máj. 5. 11:50:06 | Fast company - tech