Skip to content

Ask, Attend and Answer: Exploring Question-Guided Spatial Attention for Visual Question Answering

Notifications You must be signed in to change notification settings

BasselAli1/Ask_Attend_and_Answer

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

29 Commits
 
 
 
 
 
 

Repository files navigation

Ask_Attend_and_Answer

Ask, Attend and Answer: Exploring Question-Guided Spatial Attention for Visual Question Answering

Code

Instructions for training and testing the "SMem-VQA Two-Hop" model:

  1. Download the provided caffe folder and install caffe following the instructions in http://caffe.berkeleyvision.org/installation.html .

  2. Download MSCOCO images, and VQA annotations and questions:

    cd example/data/

    ./get_image.sh

  3. Generate the hdf5 data for training and testing:

    cd example/

    python ./data/generate_h5_data/generate_h5_data.py

  4. Train the model:

    cd example/

    run ./train/train_mm.sh

  5. Model trained on VQA dataset: SMem-VQA

  6. Predict the answers for the images and questions in VQA test-dev dataset:

    cd example/

    python ./prediction/predict_json.py

Citation

@article{xu2015ask,
    Author = {Xu, Huijuan and Saenko, Kate},
    Title = {Ask, Attend and Answer: Exploring Question-Guided Spatial Attention for Visual Question Answering},
    Journal = {arXiv preprint arXiv:1511.05234},
    Year = {2015}
}

About

Ask, Attend and Answer: Exploring Question-Guided Spatial Attention for Visual Question Answering

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • C++ 56.6%
  • Makefile 27.2%
  • Python 6.9%
  • Cuda 3.4%
  • CMake 2.3%
  • Protocol Buffer 1.9%
  • Other 1.7%