Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Reproducing CLIP score in the paper #12

Open
kkjh0723 opened this issue Jun 11, 2023 · 3 comments
Open

Reproducing CLIP score in the paper #12

kkjh0723 opened this issue Jun 11, 2023 · 3 comments

Comments

@kkjh0723
Copy link

Hi,

Thanks for opening the source code.
I'm trying to reproduce the scores for CLIP in the paper but fail to reproduce it.
I use the sample config file by changing MODE_NAME to CLIP (ViT-L/14).
I evaluate all the datasets in the corpus then average the final accuracy.
I got the following score which is quite different from the paper,

Object: 0.8205209550766983
Attribute: 0.6806109948697314
Relation: 0.67975

How can I reproduce the scores in the paper?

@ayushchakravarthy
Copy link

Hi, @kkjh0723

Did you have to make any changes to the code in order to get it working? I am also trying to replicate the CLIP result but unable to do so.

Thanks!

@kkjh0723
Copy link
Author

@ayushchakravarthy ,
If I remember correctly, there were some minor changes required to run CLIP.

In the following lines,
I changed result_tmp[i][0][1] to result_tmp[i][0][0] and result_tmp[i][1][1] to result_tmp[i][1][0].

Also, in this lines, I changed it as following,

sample_t = random.sample(sample_true,self.sample_num if len(sample_true)>self.sample_num else len(sample_true))
sample_f = random.sample(sample_false,self.sample_num if len(sample_false)>self.sample_num else len(sample_false))

@feilvvl
Copy link

feilvvl commented Oct 24, 2023

Hi,@kkjh0723
Have you reproduced the results of this work? I have tried many times, but the end result is not satisfactory.
I used the CLIP(ViT-B/32) as my model. And I select the "ITM" task to test.
For the final average scores,
Attribute : 68.6477405706409
Relation : 74.7221415628598
Object : 89.4515112110188
The result is much higher than the paper. So I'd like to know how much data you used, since your results don't vary that much.
Thank you!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants