University of Amsterdam and Renmin University at TRECVID 2017: Searching Video, Detecting Events and Describing Video

01/15/2020
by   Noureldien Hussein, et al.
0

In this paper, we summarize our TRECVID 2017 video recognition and retrieval experiments. We participated in three tasks: video search, event detection and video description. For both video search and event detection, we explore semantic representations based on VideoStory and an ImageNet Shuffle, which thrive well in few-example regimes. For the video description task, we experiment with a deep network that predicts a visual representation from a natural language description with Word2VisualVec, and use this space for the sentence matching. For generative description we enhance a neural image captioning model with Early Embedding and Late Reranking. The 2017 edition of the TRECVID benchmark has been a fruitful participation for our joint-team, resulting in the best overall result for video search and event detection as well as the runner-up position for video description.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset
Success!
Error Icon An error occurred

Sign in with Google

×

Use your Google Account to sign in to DeepAI

×

Consider DeepAI Pro