Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Can you explain the discrepancy in reported kNN Top-1 between POA and DINOv2-L? #1

Open
mranzinger opened this issue Nov 6, 2024 · 3 comments

Comments

@mranzinger
Copy link

The first image is your Table 1, and the second is DINOv2's Table 4.

Screen Shot 2024-11-06 at 12 20 59 PM Screen Shot 2024-11-06 at 12 21 21 PM

The difference between 82.0% as reported in your paper versus 83.5% in DINOv2's is quite large. It's also apparent that DINOv2-L is a patch-size-14 model, not a 16. Are you accounting for this difference by resizing DINO's patch projector? Or just letting the different models have a different number of patches? Or are you feeding them different image sizes?

@Qichuzyy
Copy link
Collaborator

Qichuzyy commented Dec 3, 2024

Here are two primary reasons for the observed performance difference:

  1. Difference of model configuration: We adjusted the model configuration for a fair comparison with most self-supervised learning methods by setting the patch size to 16. However, it is generally observed that a smaller patch size, such as 14, tends to deliver better performance than 16.
  2. Difference of pre-training methods and data: The original DINOv2 paper utilized an unreleased dataset, whereas we used the standard ImageNet-1k dataset for our reproduction. Furthermore, the ViT-L model in the original DINOv2 paper was distilled from a pre-trained ViT-H model, while we trained ViT-L model from scratch.

@mranzinger
Copy link
Author

Okay, so for your 82.0% result, you trained a new ViT-L/16 DINOv2 from scratch on ImageNet-1k, right?

@Qichuzyy
Copy link
Collaborator

Qichuzyy commented Dec 5, 2024

yes~

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants