Accurate perception of audiovisual stimuli depends crucially on the spatial and temporal properties of each sensory component, with multisensory enhancement only occurring if those components are presented in spatiotemporal congruency. Although spatial localization and temporal detection of audiovisual signals have each been extensively studied, the neural mechanisms underlying their joint influence, particularly in spatiotemporally misaligned contexts, remain poorly understood. Moreover, empirical dissection of their respective contributions to behavioral outcomes proves challenging when spatial and temporal disparities are introduced concurrently. Here, we sought to elucidate the mutual interaction of temporal and spatial offsets on the neural encoding of audiovisual stimuli. To this end, we developed a biologically inspired neurocomputational model that reproduces behavioral evidence of perceptual phenomena observed in audiovisual tasks, i.e., the modality switch effect (temporal realm) and the ventriloquist effect (spatial realm). Tested against the race model, our network successfully simulates multisensory enhancement in reaction times due to the concurrent presentation of cross-modal stimuli. Further investigation on the mechanisms implemented in the network upheld the centrality of cross-sensory inhibition in explaining modality switch effects and of cross-modal and lateral intra-area connections in regulating the evolution of these effects in space. Finally, the model predicts an amelioration in temporal detection of different modality stimuli with increasing between-stimuli eccentricity and indicates a plausible reduction in auditory localization bias for increasing interstimulus interval between spatially disparate cues. Our findings provide novel insights into the neural computations underlying audiovisual perception and offer a comprehensive predictive framework to guide future experimental investigations of multisensory integration.

Cuppini, C., Di Rosa, E.F., Astolfi, L., Monti, M. (2025). Unraveling Audiovisual Perception Across Space and Time: A Neuroinspired Computational Architecture. EJN. EUROPEAN JOURNAL OF NEUROSCIENCE, 62(3), 1-17 [10.1111/ejn.70217].

Unraveling Audiovisual Perception Across Space and Time: A Neuroinspired Computational Architecture

Cuppini Cristiano
Primo
;
Di Rosa Eleonore Federica
Secondo
;
Monti Melissa
Ultimo
2025

Abstract

Accurate perception of audiovisual stimuli depends crucially on the spatial and temporal properties of each sensory component, with multisensory enhancement only occurring if those components are presented in spatiotemporal congruency. Although spatial localization and temporal detection of audiovisual signals have each been extensively studied, the neural mechanisms underlying their joint influence, particularly in spatiotemporally misaligned contexts, remain poorly understood. Moreover, empirical dissection of their respective contributions to behavioral outcomes proves challenging when spatial and temporal disparities are introduced concurrently. Here, we sought to elucidate the mutual interaction of temporal and spatial offsets on the neural encoding of audiovisual stimuli. To this end, we developed a biologically inspired neurocomputational model that reproduces behavioral evidence of perceptual phenomena observed in audiovisual tasks, i.e., the modality switch effect (temporal realm) and the ventriloquist effect (spatial realm). Tested against the race model, our network successfully simulates multisensory enhancement in reaction times due to the concurrent presentation of cross-modal stimuli. Further investigation on the mechanisms implemented in the network upheld the centrality of cross-sensory inhibition in explaining modality switch effects and of cross-modal and lateral intra-area connections in regulating the evolution of these effects in space. Finally, the model predicts an amelioration in temporal detection of different modality stimuli with increasing between-stimuli eccentricity and indicates a plausible reduction in auditory localization bias for increasing interstimulus interval between spatially disparate cues. Our findings provide novel insights into the neural computations underlying audiovisual perception and offer a comprehensive predictive framework to guide future experimental investigations of multisensory integration.
2025
Cuppini, C., Di Rosa, E.F., Astolfi, L., Monti, M. (2025). Unraveling Audiovisual Perception Across Space and Time: A Neuroinspired Computational Architecture. EJN. EUROPEAN JOURNAL OF NEUROSCIENCE, 62(3), 1-17 [10.1111/ejn.70217].
Cuppini, Cristiano; Di Rosa, Eleonore Federica; Astolfi, Laura; Monti, Melissa
File in questo prodotto:
File Dimensione Formato  
Eur J of Neuroscience - 2025 - Cuppini - Unraveling Audiovisual Perception Across Space and Time A Neuroinspired.pdf

accesso aperto

Tipo: Versione (PDF) editoriale / Version Of Record
Licenza: Licenza per Accesso Aperto. Creative Commons Attribuzione (CCBY)
Dimensione 4.2 MB
Formato Adobe PDF
4.2 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11585/1051239
Citazioni
  • ???jsp.display-item.citation.pmc??? 1
  • Scopus 0
  • ???jsp.display-item.citation.isi??? 0
social impact