Kayser, C. and Shams, L. (2015) Multisensory causal inference in the brain. PLoS Biology, 13(2), e1002075. (doi: 10.1371/journal.pbio.1002075) (PMID:25710476) (PMCID:PMC4339834)
|
Text
104784.pdf - Published Version Available under License Creative Commons Attribution. 516kB |
Abstract
At any given moment, our brain processes multiple inputs from its different sensory modalities (vision, hearing, touch, etc.). In deciphering this array of sensory information, the brain has to solve two problems: (1) which of the inputs originate from the same object and should be integrated and (2) for the sensations originating from the same object, how best to integrate them. Recent behavioural studies suggest that the human brain solves these problems using optimal probabilistic inference, known as Bayesian causal inference. However, how and where the underlying computations are carried out in the brain have remained unknown. By combining neuroimaging-based decoding techniques and computational modelling of behavioural data, a new study now sheds light on how multisensory causal inference maps onto specific brain areas. The results suggest that the complexity of neural computations increases along the visual hierarchy and link specific components of the causal inference process with specific visual and parietal regions.
Item Type: | Articles |
---|---|
Status: | Published |
Refereed: | Yes |
Glasgow Author(s) Enlighten ID: | Kayser, Professor Christoph |
Authors: | Kayser, C., and Shams, L. |
College/School: | College of Medical Veterinary and Life Sciences > School of Psychology & Neuroscience |
Journal Name: | PLoS Biology |
Publisher: | Public Library of Science |
ISSN: | 1544-9173 |
ISSN (Online): | 1545-7885 |
Copyright Holders: | Copyright © 2015 The Authors |
First Published: | First published in PLoS Biology 13(2):e1002075 |
Publisher Policy: | Reproduced under a Creative Commons License |
University Staff: Request a correction | Enlighten Editors: Update this record