Skip to content

Unifying the Video and Question Attentions for Open-Ended Video Question Answering

Notifications You must be signed in to change notification settings

ZJULearning/videoqa

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

12 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Unifying the Video and Question Attentions for Open-Ended Video Question Answering

Table of Contents

Introduction

videoqa is the dataset and the algorithms used in Unifying the Video and Question Attentions for Open-Ended Video Question Answering

Datasets

  • file_map: contains the Tumblr urls of the videos
  • QA: contains the question-answer pairs
  • Split: contains the dataset split in the paper

Methods

Compared Algorithms

Results

ex1

  • Question: What is a boy combing his hair with?
  • Groundtruth: with his fingers
  • Prediction: with his hands

ex2

  • Question: What runs up a fence?
  • Groundtruth: a cat
  • Prediction: a cat

ex3

  • Question: What is a young girl in a car adjusting?
  • Groundtruth: her dark glasses
  • Prediction: her hair

Dependency

Usage

python main.py

Reference

If you use the code or our dataset, please cite our paper

@article{xue2017unifying,

title={Unifying the Video and Question Attentions for Open-Ended Video Question Answering},

author={Xue, Hongyang and Zhao, Zhou and Cai, Deng},

journal={IEEE Transactions on Image Processing},

year={2017},

publisher={IEEE}

}

About

Unifying the Video and Question Attentions for Open-Ended Video Question Answering

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages