Implementation of VisionLLaMA from the paper: "VisionLLaMA: A Unified LLaMA Interface for Vision Tasks" in PyTorch and Zeta
-
Updated
May 18, 2024 - Python
Implementation of VisionLLaMA from the paper: "VisionLLaMA: A Unified LLaMA Interface for Vision Tasks" in PyTorch and Zeta
This is a series of computer vision foundational projects for anyone diving into the field must tackle.
Implementation of Midas from [Towards Robust Monocular Depth Estimation] in Pytorch and Zeta
A framework to compute threshold sensitivity of deep networks to visual stimuli.
Vision-based swarms in the Presence of Occlusions
Testing the Moondream tiny vision model
In This repo i FineTuned a Pretrained ResNet18 model from PyTorch library
building AVA from ex-machina; a lightweight multi-modal system from scratch, just for learning & experimentation
Add a description, image, and links to the vision-models topic page so that developers can more easily learn about it.
To associate your repository with the vision-models topic, visit your repo's landing page and select "manage topics."