Investigative Ophthalmology & Visual Science Cover Image for Volume 65, Issue 7
June 2024
Volume 65, Issue 7
Open Access
ARVO Annual Meeting Abstract  |   June 2024
Modelling vision using deep learning techniques
Author Affiliations & Notes
  • Yalda Mohsenzadeh
    Brain and Mind Institute, Western University, London, Ontario, Canada
    Computer Science, Western University, London, Ontario, Canada
  • Footnotes
    Commercial Relationships   Yalda Mohsenzadeh None
  • Footnotes
    Support  Western BrainsCAN through CFREF, Vector Institute for AI research grant support
Investigative Ophthalmology & Visual Science June 2024, Vol.65, 1528. doi:
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Yalda Mohsenzadeh; Modelling vision using deep learning techniques. Invest. Ophthalmol. Vis. Sci. 2024;65(7):1528.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Presentation Description : Visual recognition is a fundamental function of the human brain, relying on a cascade of neural processes to transform low level inputs into semantic content. Despite significant advances in characterizing the locus and function of key perceptual cortical regions, integrating the temporal and spatial dynamics of this processing stream has been a challenge. In this talk, I will present a series of works which address this challenge by showing how the combination of MEG (or EEG), functional MRI measurements, representational geometry and deep neural networks can give new insights into visual processes in the human brain. First, I will present a novel method to characterize the interplay of feedforward and feedback mechanisms along the human ventral visual stream, and suggest how recurrent artificial neural networks can better explain the neural data in challenging visual tasks. Second, I will show how a deep generative adversarial autoencoder reveals dynamics of neural feedback processes as reconstructing low level visual information from high level latent representations. Finally, I will present the results of our latest study showing how recent developments in Large language models are the best to explain cross-modal perceptual similarities across vision and language.

This abstract was presented at the 2024 ARVO Annual Meeting, held in Seattle, WA, May 5-9, 2024.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×