August 2023
Volume 23, Issue 9
Open Access
Vision Sciences Society Annual Meeting Abstract  |   August 2023
Cortically motivated recurrence enables visual task extrapolation
Author Affiliations & Notes
  • Vijay Veerabadran
    University of California - San Diego
  • Yuan Tang
    University of California - San Diego
  • Ritik Raina
    University of California - San Diego
  • Virginia de Sa
    University of California - San Diego
  • Footnotes
    Acknowledgements  NSF:CRCNS 2208362, Kavli Institute of Brain and Mind, Sony Research Award program from Sony Research
Journal of Vision August 2023, Vol.23, 4684. doi:https://doi.org/10.1167/jov.23.9.4684
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Vijay Veerabadran, Yuan Tang, Ritik Raina, Virginia de Sa; Cortically motivated recurrence enables visual task extrapolation. Journal of Vision 2023;23(9):4684. https://doi.org/10.1167/jov.23.9.4684.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Feedforward deep neural networks have become the standard class of models in computer vision. Yet, they possess a striking difference relative to their biological counterparts which predominantly perform “recurrent” computations. Why do biological neurons evolve to employ recurrence pervasively? In this work, we show that on challenging visual tasks requiring the integration of long-range spatial dependencies, a recurrent network is able to flexibly adapt its computational budget during inference and generalize within-task across difficulties. We contribute a recurrent module we call LocRNN that is based on a prior computational model of biological vision using local recurrent intracortical connections with interneurons (Li, Z., 1998. A neural model of contour integration in the primary visual cortex. Neural Computation, 10(4), pp.903-940). LocRNN learns highly accurate solutions to the challenging visual context integration problems of Mazes and PathFinder that we use here, achieving the overall best performance across the two tasks with 3 difficulty levels each. More importantly, it is able to flexibly use less or more recurrent iterations during inference to zero-shot generalize to less- and more difficult instantiations of each task without requiring extra training data. Our observed extrapolation performance gains lie in the range of 15% to 40% on Mazes and PathFinder by varying the number of recurrent iterations, a potential functional advantage of recurrence that biological visual systems capitalize on. Our ablation study of LocRNN highlights the fundamental importance of interneurons, piecewise linear activation functions, and recurrent gating. Our work encourages further study of the role of recurrence as an important biological mechanism underlying domain generalization and task extrapolation.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×