Encoding Concept Prototypes for Video Event Detection and Summarization

Open Access
Authors
  • S.F. Chang
Publication date 2015
Book title ICMR'15: proceedings of the 2015 ACM International Conference on Multimedia Retrieval: June 23-26, 2015, Shanghai, China
ISBN
  • 9781450332743
Event 2015 ACM International Conference on Multimedia Retrieval
Pages (from-to) 123-130
Publisher New York, NY: Association for Computing Machinery
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
This paper proposes a new semantic video representation for few and zero example event detection and unsupervised video event summarization. Different from existing works, which obtain a semantic representation by training concepts over images or entire video clips, we propose an algorithm that learns a set of relevant frames as the concept prototypes from web video examples, without the need for frame-level annotations, and use them for representing an event video. We formulate the problem of learning the concept prototypes as seeking the frames closest to the densest region in the feature space of video frames from both positive and negative training videos of a target concept. We study the behavior of our video event representation based on concept prototypes by performing three experiments on challenging web videos from the TRECVID 2013 multimedia event detection task and the MED-summaries dataset. Our experiments establish that i) Event detection accuracy increases when mapping each video into concept prototype space. ii) Zero-example event detection increases by analyzing each frame of a video individually in concept prototype space, rather than considering the holistic videos. iii) Unsupervised video event summarization using concept prototypes is more accurate than using video-level concept detectors.
Document type Conference contribution
Language English
Published at https://doi.org/10.1145/2671188.2749402
Downloads
2671188.2749402 (Final published version)
Permalink to this page
Back