awesome grounding: A curated list of research papers in visual grounding
-
Updated
Apr 9, 2023
awesome grounding: A curated list of research papers in visual grounding
The Cradle framework is a first attempt at General Computer Control (GCC). Cradle supports agents to ace any computer task by enabling strong reasoning abilities, self-improvment, and skill curation, in a standardized general environment with minimal requirements.
Grounded Multimodal Large Language Model with Localized Visual Tokenization
CLIPort: What and Where Pathways for Robotic Manipulation
Code and data for "Lumos: Learning Agents with Unified Data, Modular Design, and Open-Source LLMs"
CALVIN - A benchmark for Language-Conditioned Policy Learning for Long-Horizon Robot Manipulation Tasks
PG-Video-LLaVA: Pixel Grounding in Large Multimodal Video Models
We perform functional grounding of LLMs' knowledge in BabyAI-Text
Official implementation of ICCV19 oral paper Zero-Shot grounding of Objects from Natural Language Queries (https://arxiv.org/abs/1908.07129)
[CVPR20] Video Object Grounding using Semantic Roles in Language Description (https://arxiv.org/abs/2003.10606)
Hierarchical Universal Language Conditioned Policies
[CVPR21] Visual Semantic Role Labeling for Video Understanding (https://arxiv.org/abs/2104.00990)
[Paper][AAAI 2023] DUET: Cross-modal Semantic Grounding for Contrastive Zero-shot Learning
[ICRA2023] Grounding Language with Visual Affordances over Unstructured Data
Code for CVPR'18 "Grounding Referring Expressions in Images by Variational Context"
Visual Grounding of Referring Expressions for Human-Robot Interaction
This is the official implementation for our paper;"LAR:Look Around and Refer".
[CVPR 2024] Code for "Improved Visual Grounding through Self-Consistent Explanations".
Add a description, image, and links to the grounding topic page so that developers can more easily learn about it.
To associate your repository with the grounding topic, visit your repo's landing page and select "manage topics."