Analysis of Image Captioning Models
-
Updated
Aug 6, 2017 - Python
Analysis of Image Captioning Models
CCU Computer Vision final project
Create a YOLO-format subset of the COCO dataset
A simple Python API (built on top of TensorFlow) for neural image captioning with MSCOCO data.
Object Detection Dataset Format Converter
A python based tool for looking things up in coco
Object-Detection API using MSCOCO dataset & using customized dataset from tensorflow
COCO-Stuff dataset for huggingface datasets
Image captioning with pretrained encoder on MSCOCO
Code Repository for "A New Unified Method for Detecting Text from Marathon Runners and Sports Players in Video" [Pattern Recognition, Elsevier 2020]
NLP - descriptive statistics of COCO annotations via Python COCO-API
Karpathy Splits json files for image captioning
Augment the MS COCO training set while training NIC
Mixed vision-language Attention Model that gets better by making mistakes
Microsoft COCO: Common Objects in Context for huggingface datasets
An end-to-end vision and language model incorporating explicit knowledge graphs and OOD-detection.
Caption generation from images using topics as additional guiding inputs.
MS COCO captions in Arabic
Image caption generation using GRU-based attention mechanism
Add a description, image, and links to the mscoco-dataset topic page so that developers can more easily learn about it.
To associate your repository with the mscoco-dataset topic, visit your repo's landing page and select "manage topics."