21
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Peri-personal space as a prior in coupling visual and proprioceptive signals

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          It has been suggested that the integration of multiple body-related sources of information within the peri-personal space (PPS) scaffolds body ownership. However, a normative computational framework detailing the functional role of PPS is still missing. Here we cast PPS as a visuo-proprioceptive Bayesian inference problem whereby objects we see in our environment are more likely to engender sensations as they come near to the body. We propose that PPS is the reflection of such an increased a priori probability of visuo-proprioceptive coupling that surrounds the body. To test this prediction, we immersed participants in a highly realistic virtual reality (VR) simulation of their right arm and surrounding environment. We asked participants to perform target-directed reaches toward visual, proprioceptive, and visuo-proprioceptive targets while visually displaying their reaching arm (body visible condition) or not (body invisible condition). Reach end-points are analyzed in light of the coupling prior framework, where the extension of PPS is taken to be represented by the spatial dispersion of the coupling prior between visual and proprioceptive estimates of arm location. Results demonstrate that if the body is not visible, the spatial dispersion of the visuo-proprioceptive coupling relaxes, whereas the strength of coupling remains stable. By demonstrating a distance-dependent alteration in visual and proprioceptive localization attractive pull toward one another (stronger pull at small spatial discrepancies) when the body is rendered invisible – an effect that is well accounted for by the visuo-proprioceptive coupling prior – the results suggest that the visible body grounds visuo-proprioceptive coupling preferentially in the near vs. far space.

          Related collections

          Most cited references58

          • Record: found
          • Abstract: found
          • Article: not found

          The ventriloquist effect results from near-optimal bimodal integration.

          Ventriloquism is the ancient art of making one's voice appear to come from elsewhere, an art exploited by the Greek and Roman oracles, and possibly earlier. We regularly experience the effect when watching television and movies, where the voices seem to emanate from the actors' lips rather than from the actual sound source. Originally, ventriloquism was explained by performers projecting sound to their puppets by special techniques, but more recently it is assumed that ventriloquism results from vision "capturing" sound. In this study we investigate spatial localization of audio-visual stimuli. When visual localization is good, vision does indeed dominate and capture sound. However, for severely blurred visual stimuli (that are poorly localized), the reverse holds: sound captures vision. For less blurred stimuli, neither sense dominates and perception follows the mean position. Precision of bimodal localization is usually better than either the visual or the auditory unimodal presentation. All the results are well explained not by one sense capturing the other, but by a simple model of optimal combination of visual and auditory information.
            Bookmark
            • Record: found
            • Abstract: found
            • Article: not found

            On the other hand: dummy hands and peripersonal space.

            Where are my hands? The brain can answer this question using sensory information arising from vision, proprioception, or touch. Other sources of information about the position of our hands can be derived from multisensory interactions (or potential interactions) with our close environment, such as when we grasp or avoid objects. The pioneering study of multisensory representations of peripersonal space was published in Behavioural Brain Research almost 30 years ago [Rizzolatti G, Scandolara C, Matelli M, Gentilucci M. Afferent properties of periarcuate neurons in macaque monkeys. II. Visual responses. Behav Brain Res 1981;2:147-63]. More recently, neurophysiological, neuroimaging, neuropsychological, and behavioural studies have contributed a wealth of evidence concerning hand-centred representations of objects in peripersonal space. This evidence is examined here in detail. In particular, we focus on the use of artificial dummy hands as powerful instruments to manipulate the brain's representation of hand position, peripersonal space, and of hand ownership. We also review recent studies of the 'rubber hand illusion' and related phenomena, such as the visual capture of touch, and the recalibration of hand position sense, and discuss their findings in the light of research on peripersonal space. Finally, we propose a simple model that situates the 'rubber hand illusion' in the neurophysiological framework of multisensory hand-centred representations of space.
              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              Integration of proprioceptive and visual position-information: An experimentally supported model.

              To localize one's hand, i.e., to find out its position with respect to the body, humans may use proprioceptive information or visual information or both. It is still not known how the CNS combines simultaneous proprioceptive and visual information. In this study, we investigate in what position in a horizontal plane a hand is localized on the basis of simultaneous proprioceptive and visual information and compare this to the positions in which it is localized on the basis of proprioception only and vision only. Seated at a table, subjects matched target positions on the table top with their unseen left hand under the table. The experiment consisted of three series. In each of these series, the target positions were presented in three conditions: by vision only, by proprioception only, or by both vision and proprioception. In one of the three series, the visual information was veridical. In the other two, it was modified by prisms that displaced the visual field to the left and to the right, respectively. The results show that the mean of the positions indicated in the condition with both vision and proprioception generally lies off the straight line through the means of the other two conditions. In most cases the mean lies on the side predicted by a model describing the integration of multisensory information. According to this model, the visual information and the proprioceptive information are weighted with direction-dependent weights, the weights being related to the direction-dependent precision of the information in such a way that the available information is used very efficiently. Because the proposed model also can explain the unexpectedly small sizes of the variable errors in the localization of a seen hand that were reported earlier, there is strong evidence to support this model. The results imply that the CNS has knowledge about the direction-dependent precision of the proprioceptive and visual information.
                Bookmark

                Author and article information

                Contributors
                max.diluca@oculus.com
                Journal
                Sci Rep
                Sci Rep
                Scientific Reports
                Nature Publishing Group UK (London )
                2045-2322
                25 October 2018
                25 October 2018
                2018
                : 8
                : 15819
                Affiliations
                [1 ]ISNI 0000 0004 0615 529X, GRID grid.453567.6, Oculus Research, , Facebook Inc., ; Redmond, WA USA
                [2 ]ISNI 0000 0001 2264 7217, GRID grid.152326.1, Vanderbilt Brain Institute, , Vanderbilt University, ; Nashville, TN USA
                [3 ]ISNI 0000 0000 9632 6718, GRID grid.19006.3e, Department of Psychology, , University of California Los Angeles, ; Los Angeles, CA USA
                [4 ]ISNI 0000 0004 1936 7486, GRID grid.6572.6, Centre for Computational Neuroscience and Cognitive Robotics, , University of Birmingham, ; Birmingham, UK
                Article
                33961
                10.1038/s41598-018-33961-3
                6202371
                30361477
                e4ff83bc-9d45-4022-a4fe-a03667ab8e66
                © The Author(s) 2018

                Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.

                History
                : 24 April 2018
                : 7 October 2018
                Categories
                Article
                Custom metadata
                © The Author(s) 2018

                Uncategorized
                Uncategorized

                Comments

                Comment on this article