Attention guides our gaze to fixate the proper location of the scene and holds it in that location for the deserved amount of time given current processing demands, before shifting to the next one. As such, gaze deployment crucially is a temporal process. Existing computational models have made significant strides in predicting spatial aspects of observer's visual scanpaths (where to look), while often putting on the background the temporal facet of attention dynamics (when). In this paper we present TPP-Gaze, a novel and principled approach to model scanpath dynamics based on Neural Temporal Point Process (TPP), that jointly learns the temporal dynamics of fixations position and duration, integrating deep learning methodologies with point process theory. We conduct extensive experiments across five publicly available datasets. Our results show the overall superior performance of the proposed model compared to state-of-the-art approaches.

TPP-Gaze: Modelling Gaze Dynamics in Space and Time with Neural Temporal Point Processes / D'Amelio, Alessandro; Cartella, Giuseppe; Cuculo, Vittorio; Lucchi, Manuele; Cornia, Marcella; Cucchiara, Rita; Boccignone, Giuseppe. - (2025), pp. 8786-8795. ( 2025 IEEE/CVF Winter Conference on Applications of Computer Vision, WACV 2025 Tucson, Arizona February 28 - March 4) [10.1109/WACV61041.2025.00851].

TPP-Gaze: Modelling Gaze Dynamics in Space and Time with Neural Temporal Point Processes

Giuseppe Cartella;Vittorio Cuculo;Marcella Cornia;Rita Cucchiara;
2025

Abstract

Attention guides our gaze to fixate the proper location of the scene and holds it in that location for the deserved amount of time given current processing demands, before shifting to the next one. As such, gaze deployment crucially is a temporal process. Existing computational models have made significant strides in predicting spatial aspects of observer's visual scanpaths (where to look), while often putting on the background the temporal facet of attention dynamics (when). In this paper we present TPP-Gaze, a novel and principled approach to model scanpath dynamics based on Neural Temporal Point Process (TPP), that jointly learns the temporal dynamics of fixations position and duration, integrating deep learning methodologies with point process theory. We conduct extensive experiments across five publicly available datasets. Our results show the overall superior performance of the proposed model compared to state-of-the-art approaches.
2025
2025 IEEE/CVF Winter Conference on Applications of Computer Vision, WACV 2025
Tucson, Arizona
February 28 - March 4
8786
8795
D'Amelio, Alessandro; Cartella, Giuseppe; Cuculo, Vittorio; Lucchi, Manuele; Cornia, Marcella; Cucchiara, Rita; Boccignone, Giuseppe
TPP-Gaze: Modelling Gaze Dynamics in Space and Time with Neural Temporal Point Processes / D'Amelio, Alessandro; Cartella, Giuseppe; Cuculo, Vittorio; Lucchi, Manuele; Cornia, Marcella; Cucchiara, Rita; Boccignone, Giuseppe. - (2025), pp. 8786-8795. ( 2025 IEEE/CVF Winter Conference on Applications of Computer Vision, WACV 2025 Tucson, Arizona February 28 - March 4) [10.1109/WACV61041.2025.00851].
File in questo prodotto:
File Dimensione Formato  
TPP-Gaze_Modelling_Gaze_Dynamics_in_Space_and_Time_with_Neural_Temporal_Point_Processes.pdf

Accesso riservato

Tipologia: VOR - Versione pubblicata dall'editore
Dimensione 4.08 MB
Formato Adobe PDF
4.08 MB Adobe PDF   Visualizza/Apri   Richiedi una copia
Pubblicazioni consigliate

Licenza Creative Commons
I metadati presenti in IRIS UNIMORE sono rilasciati con licenza Creative Commons CC0 1.0 Universal, mentre i file delle pubblicazioni sono rilasciati con licenza Attribuzione 4.0 Internazionale (CC BY 4.0), salvo diversa indicazione.
In caso di violazione di copyright, contattare Supporto Iris

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11380/1362546
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 1
  • ???jsp.display-item.citation.isi??? 0
social impact