Skip to content

ramprasadkillari/Image-Captioning

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Image Captioning

Deep Learning model to generate a caption for a given image.
See report.pdf for more details about the architecture of model.
Used VGG16 and InceptionV3 models to extract features from the Image.

Datasets:

To just visualise the results and model's output, download the Caption folder which contains pre-trained model. For training the model from scratch download the Flickr Datasets.
Flickr 8k
Flickr 30k

Usage:

Captioning

cd Caption/
python3 caption.py dog.jpg

Training

cd Training/
python3 features.py #Saves the features into features.pkl file
python3 train.py

Output:

alt text

Note:

First time you train this model, Keras will download the model weights from the Internet, which are about 500 Megabytes. This may take a few minutes.
If you don't want to train the model, Just download the Caption Folder, thus saving time.

References:

machinelearningmastery
Andrej Karpathy

About

Deep Learning model to caption an image. [Course Project, CS 337( 2019-1 ) - IIT Bombay]

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages