Skip to content

Commit

Permalink
EVA-CLIP+ (WIP)
Browse files Browse the repository at this point in the history
  • Loading branch information
Yuxin-CV committed Feb 3, 2023
1 parent 6793e2a commit 4f3f50a
Showing 1 changed file with 3 additions and 1 deletion.
4 changes: 3 additions & 1 deletion clip/benchmark.md
Original file line number Diff line number Diff line change
Expand Up @@ -948,4 +948,6 @@ The zero-shot retrieval performance of EVA-CLIP is relatively inferior to the Op
- The size / capacity of the language tower in EVA-CLIP is much smaller / weaker than Open CLIP-H and Open CLIP-g, *i.e.*, `124M` *v.s.* `354M`. Meanwhile, retrieval tasks depend more on the capacity of the language branch compared with classification tasks.
- Retrieval tasks seem benefit more from the training dataset size (LAION-2B used by Open CLIP), while we only leverage LAION-400M for EVA-CLIP training.

Nevertheless, it is hard to make a head-to-head comparison between different CLIP models. In the future, we will further scale up the language encoder & training data to improve the retrieval performance.
Nevertheless, it is hard to make a head-to-head comparison between different CLIP models. In the future, we will further scale up the language encoder & training data to improve the retrieval performance.

- **Updates (Feb, 2023)**: We are training an improved version of EVA-CLIP+ (WIP), now achieving ~79.5 zero-shot top-1 accuracy on IN-1K, and outperforming the prev. best CLIP by ~0.5\% in zero-shot retrieval. We will update the details soon and release all suites of EVA-CLIP+ in the future. Please stay tuned.

0 comments on commit 4f3f50a

Please sign in to comment.