People Pay Video Game Using Only Direct Brain Stimulation

Tuesday, December 6, 2016

People Pay Video Game Using Only Direct Brain Stimulation


Direct Brain Stimulation


Scientists have published the first demonstration of humans playing a simple, two-dimensional computer game using only input from direct brain stimulation—without relying on any usual sensory cues from sight, hearing or touch.

University of Washington researchers have taken a first step in showing how humans can interact with virtual realities via direct brain stimulation.

In a paper published online in Frontiers in Robotics and AI, they describe their demonstration of humans playing a simple, two-dimensional computer game using only input from direct brain stimulation — without relying on any usual sensory cues from sight, hearing or touch.

In the game, the subjects had to navigate 21 different mazes, with two choices to move forward or down based on whether they sensed a visual stimulation artifact called a phosphene, which are perceived as blobs or bars of light. To signal which direction to move, the researchers generated a phosphene through transcranial magnetic stimulation, a well-known technique that uses a magnetic coil placed near the skull to directly and noninvasively stimulate a specific area of the brain.

“The way virtual reality is done these days is through displays, headsets and goggles, but ultimately your brain is what creates your reality,” said senior author Rajesh Rao, UW professor of Computer Science & Engineering and director of the Center for Sensorimotor Neural Engineering.

“The fundamental question we wanted to answer was: Can the brain make use of artificial information that it’s never seen before that is delivered directly to the brain to navigate a virtual world or do useful tasks without other sensory input? And the answer is yes.”

The five test subjects made the right moves in the mazes 92 percent of the time when they received the input via direct brain stimulation, compared to 15 percent of the time when they lacked that guidance.

The absence or presence of phosphenes – visual artifacts that can be created through direct brain stimulation – told the test subjects whether to move forward or down.University of Washington

"The way virtual reality is done these days is through displays, headsets and goggles, but ultimately your brain is what creates your reality."
The simple game demonstrates one way that novel information from artificial sensors or computer-generated virtual worlds can be successfully encoded and delivered noninvasively to the human brain to solve useful tasks. It employs a technology commonly used in neuroscience to study how the brain works — transcranial magnetic stimulation — to instead convey actionable information to the brain.

The test subjects also got better at the navigation task over time, suggesting that they were able to learn to better detect the artificial stimuli.

“We’re essentially trying to give humans a sixth sense,” said lead author Darby Losey, a graduate in computer science and neurobiology who now works as a staff researcher for the Institute for Learning & Brain Sciences (I-LABS).  “So much effort in this field of neural engineering has focused on decoding information from the brain. We’re interested in how you can encode information into the brain.”

"These results suggest that humans can learn to utilize information delivered non-invasively to their brains to solve tasks that cannot be solved using their natural senses. Exploring this emerging field of human sensory augmentation, with its technological as well as ethical and social implications, remains an active area of research," conclude the researchers.

People Pay Video Game Using Only Direct Brain Stimulation

Related articles
The initial experiment used binary information — whether a phosphene was present or not — to let the game players know whether there was an obstacle in front of them in the maze. In the real world, even that type of simple input could help blind or visually impaired individuals navigate.

“The technology is not there yet — the tool we use to stimulate the brain is a bulky piece of equipment that you wouldn’t carry around with you,” said co-author Andrea Stocco, a UW assistant professor of psychology and I-LABS research scientist. “But eventually we might be able to replace the hardware with something that’s amenable to real world applications.”

The testers successfully navigated an average of 92 percent of the moves when they received input via direct brain stimulation to guide them through the experimental mazes (blue) versus only 15 percent of the steps in the control mazes when they received no such input (red mazes).University of Washington

Together with other partners from outside UW, members of the research team have co-founded Neubay, a startup company aimed at commercializing their ideas and introducing neuroscience and artificial intelligence (AI) techniques that could make virtual-reality, gaming and other applications better and more engaging.

The team is currently investigating how altering the intensity and location of direct brain stimulation can create more complex visual and other sensory perceptions which are currently difficult to replicate in augmented or virtual reality.

“We look at this as a very small step toward the grander vision of providing rich sensory input to the brain directly and noninvasively,” said Rao. “Over the long term, this could have profound implications for assisting people with sensory deficits while also paving the way for more realistic virtual reality experiences.”





SOURCE  University of Washington


By  33rd SquareEmbed



0 comments:

Post a Comment