September 2019
Volume 19, Issue 10
Open Access
Vision Sciences Society Annual Meeting Abstract  |   September 2019
Learning to Integrate Egocentric and Allocentric Information using a Goal-directed Reward Signal
Author Affiliations & Notes
  • Arthur W Juliani
    University of Oregon
  • Joseph P Yaconelli
    University of Oregon
  • Margaret E Sereno
    University of Oregon
Journal of Vision September 2019, Vol.19, 162. doi:https://doi.org/10.1167/19.10.162
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Arthur W Juliani, Joseph P Yaconelli, Margaret E Sereno; Learning to Integrate Egocentric and Allocentric Information using a Goal-directed Reward Signal. Journal of Vision 2019;19(10):162. https://doi.org/10.1167/19.10.162.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Recent work in Deep Reinforcement Learning has demonstrated the ability for a parameterized model to learn to solve complex tasks from a sparse reward signal. A consequence of this learning is often a meaningful latent representation of the observational data. The composite nature of neural networks opens the possibility of learning joint representations between not just one, but multiple sensory streams of information. In this work, we train a neural network to learn a joint spatial representation that combines separate egocentric and allocentric visual streams, corresponding to a 3D first-person view and 2D map view. We used a simple 3D environment with a goal-driven navigation task. In order to fully explore the relationship between the two information streams, we employed multiple experimental conditions where each stream contained variable amounts of relevant spatial information, specified as follows. The egocentric perspective could contain one of three levels of information (“None”, “Partial” - the goal is invisible, or “Full” - the goal is visible). Likewise, the allocentric perspective contained one of three levels of information: (“None”, “Partial” - the goal is present, but self-location is not indicated, or “Full” both the goal position and self-location are indicated). We demonstrate the novel result that a goal-driven reward signal can be used to guide the learning of a joint representation between allocen-tric and egocentric visual streams. Additionally, in the condition involving imperfect information from both streams (“Partial” - “Partial”) the network was able to learn to successfully combine the streams in a representation that contains near-perfect global self-location and orientation information, even when this information was not explicitly available in either visual stream, and allowed for near-optimal performance. We compare these learned representations to those prototypical of the mammalian “cognitive map,” as well as compare behavior results between our trained models and human participants.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×