We live in a world where our eyes and ears are almost constantly bombarded with colors, shapes, textures and noises of all types. How exactly do our brains translate these sights and sounds into meaningful images and words? At the University of California, Berkeley, two groups of scientists are finding tantalizing new answers to this question. Their remarkable successes at reconstructing what our brains see and hear offer hope for future life-changing technologies.
This video is based on:
Reconstructing Visual Experiences from Brain Activity Evoked by Natural Movies, a paper published by Shinji Nishimoto, Jack Gallant, and colleagues in the journal Current Biology.
Reconstructing Speech from Human Auditory Cortex, a paper published by Brian Pasley, Robert Knight and colleagues in the online journal PLoS Biology.
Influence of Context and Behavior on Stimulus Reconstruction From Neural Activity in Primary Auditory Cortex, a paper published by Shihab A. Shamma and colleagues in the journal Journal of Neurophysiology.
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivs 3.0 Unported License.
Use the Form Below to comment on this post.