Decoding Our Senses
We live in a world where our eyes and ears are almost constantly bombarded with colors, shapes, textures and noises of all types. How exactly do our brains translate these sights and sounds into meaningful images and words? At the University of California, Berkeley, two groups of scientists are finding tantalizing new answers to this question. Their remarkable successes at reconstructing what our brains see and hear offer hope for future life-changing technologies.
This video is based on:
“Reconstructing Visual Experiences from Brain Activity Evoked by Natural Movies,” a paper published by Jack Gallant and colleagues in the journal Current Biology.
“Reconstructing Speech from Human Auditory Cortex,” a paper published by Brian Pasley and colleagues in the online journal PLoS Biology.
“Influence of Context and Behavior on Stimulus Reconstruction From Neural Activity in Primary Auditory Cortex,” a paper published by Nima Mesgarani and colleagues in the Journal of Neurophysiology.