Multi-Task Edge Prediction in Temporally-Dynamic Video Graphs

Open Access
Authors
  • O. Ülger
  • J. Wiederer
  • M. Ghafoorian
  • V. Belagiannis
Publication date 2022
Book title The 33rd British Machine Vision Conference Proceedings
Book subtitle BMVC 2022 : 21st-24th November 2022, London, UK
Event 33rd British Machine Vision Conference
Article number 968
Number of pages 14
Publisher BMVA Press
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Graph neural networks have shown to learn effective node representations, enabling node-, link-, and graph-level inference. Conventional graph networks assume static relations between nodes, while relations between entities in a video often evolve over time, with nodes entering and exiting dynamically. In such temporally-dynamic graphs, a core problem is inferring the future state of spatio-temporal edges, which can constitute multiple types of relations. To address this problem, we propose MTD-GNN, a graph network for predicting temporally-dynamic edges for multiple types of relations. We propose a factorized spatio-temporal graph attention layer to learn dynamic node representations and present a multi-task edge prediction loss that models multiple relations simultaneously. The proposed architecture operates on top of scene graphs that we obtain from videos through object detection and spatio-temporal linking. Experimental evaluations on ActionGenome and CLEVRER show that modeling multiple relations in our temporally-dynamic graph network can be mutually beneficial, outperforming existing static and spatio-temporal graph neural networks, as well as state-of-the-art predicate classification methods. Code is available at https://github.com/ozzyou/MTD-GNN.
Document type Conference contribution
Note With supplementary poster and video
Language English
Published at https://bmvc2022.mpi-inf.mpg.de/968/
Other links https://github.com/ozzyou/MTD-GNN
Downloads
0968 (Final published version)
Supplementary materials
Permalink to this page
Back