You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Thanks for opening the source code.
I'm trying to reproduce the scores for CLIP in the paper but fail to reproduce it.
I use the sample config file by changing MODE_NAME to CLIP (ViT-L/14).
I evaluate all the datasets in the corpus then average the final accuracy.
I got the following score which is quite different from the paper,
Hi,@kkjh0723
Have you reproduced the results of this work? I have tried many times, but the end result is not satisfactory.
I used the CLIP(ViT-B/32) as my model. And I select the "ITM" task to test.
For the final average scores,
Attribute : 68.6477405706409
Relation : 74.7221415628598
Object : 89.4515112110188
The result is much higher than the paper. So I'd like to know how much data you used, since your results don't vary that much.
Thank you!
Hi,
Thanks for opening the source code.
I'm trying to reproduce the scores for CLIP in the paper but fail to reproduce it.
I use the sample config file by changing
MODE_NAME
to CLIP (ViT-L/14).I evaluate all the datasets in the corpus then average the final accuracy.
I got the following score which is quite different from the paper,
Object: 0.8205209550766983
Attribute: 0.6806109948697314
Relation: 0.67975
How can I reproduce the scores in the paper?
The text was updated successfully, but these errors were encountered: