Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

chore: polish benchmark doc #839

Merged
merged 7 commits into from
Oct 10, 2022
Merged

chore: polish benchmark doc #839

merged 7 commits into from
Oct 10, 2022

Conversation

numb3r3
Copy link
Member

@numb3r3 numb3r3 commented Oct 10, 2022

No description provided.

@codecov
Copy link

codecov bot commented Oct 10, 2022

Codecov Report

Merging #839 (4bc1f4d) into clip-benchmark (cc0e98c) will not change coverage.
The diff coverage is n/a.

@@               Coverage Diff               @@
##           clip-benchmark     #839   +/-   ##
===============================================
  Coverage           81.58%   81.58%           
===============================================
  Files                  21       21           
  Lines                1575     1575           
===============================================
  Hits                 1285     1285           
  Misses                290      290           
Flag Coverage Δ
cas 81.58% <ø> (ø)

Flags with carried forward coverage won't be shown. Click here to find out more.

Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here.

@numb3r3 numb3r3 requested review from ZiniuYu and jemmyshin October 10, 2022 05:39

Zero-shot retrieval
Zero-shot Retrieval
+++++++++++++++++++

In zero-shot retrieval benchmark, each model is evaluated on the following datasets: `COCO Caption <https://github.com/tylin/coco-caption>`_, `Flickr8k <http://hockenmaier.cs.illinois.edu/8k-pictures.html>`_ and `Flickr30k <https://shannon.cs.illinois.edu/DenotationGraph/>`_.
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
In zero-shot retrieval benchmark, each model is evaluated on the following datasets: `COCO Caption <https://github.com/tylin/coco-caption>`_, `Flickr8k <http://hockenmaier.cs.illinois.edu/8k-pictures.html>`_ and `Flickr30k <https://shannon.cs.illinois.edu/DenotationGraph/>`_.
In the zero-shot retrieval benchmark, each model is evaluated on the following datasets: `COCO Caption <https://github.com/tylin/coco-caption>`_, `Flickr8k <http://hockenmaier.cs.illinois.edu/8k-pictures.html>`_ and `Flickr30k <https://shannon.cs.illinois.edu/DenotationGraph/>`_.
The best results are highlighted in bold (higher is better).

@@ -151,7 +154,7 @@ From the table, we observe that the ViT models outperform the RN models in gener
More specifically, the ``ViT-H-14::laion2b_s32b_b79k`` model and ``ViT-g-14::laion2b_s12b_b42k`` model achieve the best and second-best results on all zero-shot retrieval tasks.
For ViT models, the results of the same base model are better on those pre-trained with larger datasets (e.g., ``ViT-B-32::openai`` vs ``ViT-B-32::laion400m_e31`` vs ``ViT-B-32::laion2b-s34b-b79k``).

Zero-shot classification
Zero-shot Classification
++++++++++++++++++++++++

In zero-shot classification benchmark, each model is evaluated on the following datasets: `ImageNetV2 <https://github.com/modestyachts/ImageNetV2>`_, `VOC2007 <http://host.robots.ox.ac.uk/pascal/VOC/voc2007/>`_ and 19 `VTAB datasets <https://github.com/google-research/task_adaptation>`_.
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
In zero-shot classification benchmark, each model is evaluated on the following datasets: `ImageNetV2 <https://github.com/modestyachts/ImageNetV2>`_, `VOC2007 <http://host.robots.ox.ac.uk/pascal/VOC/voc2007/>`_ and 19 `VTAB datasets <https://github.com/google-research/task_adaptation>`_.
In the zero-shot classification benchmark, each model is evaluated on the following datasets: `ImageNetV2 <https://github.com/modestyachts/ImageNetV2>`_, `VOC2007 <http://host.robots.ox.ac.uk/pascal/VOC/voc2007/>`_ and 19 `VTAB datasets <https://github.com/google-research/task_adaptation>`_.
The best results are highlighted in bold (higher is better).

@numb3r3 numb3r3 requested a review from ZiniuYu October 10, 2022 05:51
@github-actions
Copy link

📝 Docs are deployed on https://ft-polish-clip-benchmark--jina-docs.netlify.app 🎉

@numb3r3 numb3r3 merged commit 9839451 into clip-benchmark Oct 10, 2022
@numb3r3 numb3r3 deleted the polish-clip-benchmark branch October 10, 2022 06:06
numb3r3 added a commit that referenced this pull request Oct 10, 2022
…832)

* docs: clip benchmark on zeroshot classification and retrieval tasks

* docs: add label

* docs: introduction

* docs: open clip naming convention

* fix: typo

* docs: retrieval table

* docs: update classification

* chore: test html table

* chore: update css

* chore: test rst

* chore: test rst

* chore: test

* fix: use rst in benchmark

* fix: typo

* fix: rst

* fix: rst

* fix: subtitle

* docs: classification benchmark

* docs: highlight retrieval

* docs: highlight retireval

* docs: highlight classification

* docs: remove redundancy

* docs: add links

* fix: link

* docs: update section

* docs: datasets description

* docs: add datasets description

* docs: format

* docs: footnote

* docs: add QPS

* docs: improve conclusion

* docs: update machine config

* docs: update software version

* chore: polish benchmark doc (#839)

* chore: update benchmark intro

* chore: minor revision

* chore: minor revision

* chore: minor revision

* chore: minor revision

* chore: minor revision

* chore: minor revision

Co-authored-by: felix-wang <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants