Skip to content
master
Switch branches/tags
Go to file
Code

Latest commit

 

Git stats

Files

Permalink
Failed to load latest commit information.
Type
Name
Latest commit message
Commit time
 
 
 
 
 
 
 
 
 
 
 
 

README.md

GHA-ImageCaptioning

This is the code for our papers: Gated Hierarchical Attention for Image Captioning and CNN+CNN: Convolutional Decoders for Image Captioning. To run it you should first install Pytorch 0.3.0.

Train

  1. Download the MSCOCO2014 dataset here.
  2. Unzip the files, and you put the training and validation images in the same folder. Put captions_val2014.json file in the annotation folder.
  3. Download Karpathy's split here, and put it in the folder data/files/, then run ak_build_vocab.py in the data folder to preprocess the dataset.
  4. Download COCO evaluation metrics here. Copy all files to models/coco_eval.
  5. Let self.image_dir in train.py equal to the path of the folder in step 2. Also, you can change other parameters in the configuration.

Inference

After training you can use the inference.py to generate captions for the images in the test split. Also, you should assign the path of the image folder to self.image_dir.

About

Code for GHA (ACCV2018)

Topics

Resources

Releases

No releases published

Packages

No packages published

Languages