A Dataset and Exploration of Models for Understanding Video Data through Fill-in-the-Blank Question-Answering

التفاصيل البيبلوغرافية
العنوان: A Dataset and Exploration of Models for Understanding Video Data through Fill-in-the-Blank Question-Answering
المؤلفون: Nicolas Ballas, Chris Pal, Aaron Courville, Tegan Maharaj, Anna Rohrbach
المصدر: CVPR
بيانات النشر: IEEE, 2017.
سنة النشر: 2017
مصطلحات موضوعية: FOS: Computer and information sciences, Vocabulary, Computer science, Computer Vision and Pattern Recognition (cs.CV), media_common.quotation_subject, Computer Science - Computer Vision and Pattern Recognition, 02 engineering and technology, 010501 environmental sciences, Machine learning, computer.software_genre, 01 natural sciences, Convolutional neural network, 0202 electrical engineering, electronic engineering, information engineering, Question answering, 0105 earth and related environmental sciences, media_common, business.industry, Automatic summarization, Object detection, Visualization, 020201 artificial intelligence & image processing, Artificial intelligence, Language model, business, computer, Natural language
الوصف: While deep convolutional neural networks frequently approach or exceed human-level performance at benchmark tasks involving static images, extending this success to moving images is not straightforward. Having models which can learn to understand video is of interest for many applications, including content recommendation, prediction, summarization, event/object detection and understanding human visual perception, but many domains lack sufficient data to explore and perfect video models. In order to address the need for a simple, quantitative benchmark for developing and understanding video, we present MovieFIB, a fill-in-the-blank question-answering dataset with over 300,000 examples, based on descriptive video annotations for the visually impaired. In addition to presenting statistics and a description of the dataset, we perform a detailed analysis of 5 different models' predictions, and compare these with human performance. We investigate the relative importance of language, static (2D) visual features, and moving (3D) visual features; the effects of increasing dataset size, the number of frames sampled; and of vocabulary size. We illustrate that: this task is not solvable by a language model alone; our model combining 2D and 3D visual information indeed provides the best result; all models perform significantly worse than human-level. We provide human evaluations for responses given by different models and find that accuracy on the MovieFIB evaluation corresponds well with human judgement. We suggest avenues for improving video models, and hope that the proposed dataset can be useful for measuring and encouraging progress in this very interesting field.
URL الوصول: https://explore.openaire.eu/search/publication?articleId=doi_dedup___::bc063795248804a12526337a99f21d5c
https://doi.org/10.1109/cvpr.2017.778
حقوق: OPEN
رقم الأكسشن: edsair.doi.dedup.....bc063795248804a12526337a99f21d5c
قاعدة البيانات: OpenAIRE