August 2016
Volume 16, Issue 12
Open Access
Vision Sciences Society Annual Meeting Abstract  |   September 2016
Biophysically plausible neural model for the interaction between action observation and action execution
Author Affiliations
  • Mohammad Hovaidi Ardestani
    Section for Computational Sensomotorics, Department of Cognitive Neurology, Hertie Institute for Clinical Brain Research, and Centre for Integrative Neuroscience, University Clinic Tübingen, D-72076 Tübingen, Germany
  • Martin Giese
    Section for Computational Sensomotorics, Department of Cognitive Neurology, Hertie Institute for Clinical Brain Research, and Centre for Integrative Neuroscience, University Clinic Tübingen, D-72076 Tübingen, Germany
Journal of Vision September 2016, Vol.16, 980. doi:10.1167/16.12.980
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to Subscribers Only
      Sign In or Create an Account ×
    • Get Citation

      Mohammad Hovaidi Ardestani, Martin Giese; Biophysically plausible neural model for the interaction between action observation and action execution. Journal of Vision 2016;16(12):980. doi: 10.1167/16.12.980.

      Download citation file:


      © 2017 Association for Research in Vision and Ophthalmology.

      ×
  • Supplements
Abstract

INTRODUCTION: It has been suggested that neural representations for action perception and action execution are closely linked in the human brain. Previous studies show that the execution of motor behavior influences concurrent visual action observation, and specifically the perception of biological motion. The detailed neural mechanisms that support this interaction are unknown. METHODS: We extended a physiologically-inspired model for visual action recognition (Fleischer et. al. J. Nsc. 2013) by a neural representation of motor programs, which is coupled reciprocally with visual recognition layer that is modeled by a dynamic neural network. Visual and executed actions are represented by a localized activation maxima that propagates in the visual and in the motor representation, which are modeled by dynamic neural fields that are implemented with biophysically realistic spiking neurons. An appropriately designed coupling between both representations results in interactions between action observation and action execution that mimic experimental results on the interactions between motor execution and recognition (Christensen et al., J. Nsc. 2011; Kilner et al., 2003). RESULTS: Consistent with the experimental data, we find a facilitation of the detection of visual action patterns by concurrent motor execution if the executed motor pattern is spatio-temporally compatible with the observed pattern, and interference if it is incoherent, e.g. in presence of time delays between observed and executed actions. CONCLUSION: Simple dynamic neural networks with biophysically realistic neurons can reproduce basic signatures of perception-action coupling in behavioral experiments.

Meeting abstract presented at VSS 2016

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×