Analysis of Image Captioning Models
-
Updated
Aug 6, 2017 - Python
Analysis of Image Captioning Models
CCU Computer Vision final project
Create a YOLO-format subset of the COCO dataset
Microsoft COCO: Common Objects in Context for huggingface datasets
A simple Python API (built on top of TensorFlow) for neural image captioning with MSCOCO data.
NLP - descriptive statistics of COCO annotations via Python COCO-API
Mixed vision-language Attention Model that gets better by making mistakes
Object Detection Dataset Format Converter
A python based tool for looking things up in coco
Karpathy Splits json files for image captioning
COCO-Stuff dataset for huggingface datasets
Image captioning with pretrained encoder on MSCOCO
Augment the MS COCO training set while training NIC
Code Repository for "A New Unified Method for Detecting Text from Marathon Runners and Sports Players in Video" [Pattern Recognition, Elsevier 2020]
An end-to-end vision and language model incorporating explicit knowledge graphs and OOD-detection.
Caption generation from images using topics as additional guiding inputs.
MS COCO captions in Arabic
Object-Detection API using MSCOCO dataset & using customized dataset from tensorflow
Image caption generation using GRU-based attention mechanism
Add a description, image, and links to the mscoco-dataset topic page so that developers can more easily learn about it.
To associate your repository with the mscoco-dataset topic, visit your repo's landing page and select "manage topics."