-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Paper‘s result cannot be reproduced #12
Comments
[Setup] [Results] |
Thank you for your reply, yes, the only probelm is FID. And we found the last model's evaluation is better than the best model and very close to pretrained model, it maybe the overfitting thing. I've checked the number of image that used for evaluation, the number it's correct. And the script used for get FID is Threedfront, we did not change that, do you have any other advice for FID's problem? Still can't solve the FID problem. |
I can't tell what might be wrong here. It is strange that KID results are close but FID are not given the same inputs. I can run the evaluation script on my side if you send me an example set of 1000 synthetic layout images. |
Thank you very much for your help and patience. I just ran it again and found that there is still a problem with the FID. And the file is too large to upload on github, so I sent the livingroom synthetic layout images to your email. Hope you can help me check what is wrong with my evaluation operation. |
And we use headless rendering, I don't know if this will cause some problem. |
Thank you for your excellent work! However, when we tried to reproduce the results reported in your paper, we can't reprodece the paper's results.Here are the details of our attempt and the problem we met:
1.We did not adjust any training parameters and used the exact configurations provided in the MiDiffusion/config/ yaml files, including epoch, learning rate, and other settings. Are there any additional tricks or adjustments required during training?
2.For the PointNet feature extractor, should it be a pretrained version, or is it intended to be trained from scratch?
3.For dataset preprocessing, we directly used the files from the ThreedFront dataset. Are there any specific preprocessing steps or modifications needed that are not mentioned in the paper?
4.Even when we used the pretrained weights provided by you for evaluation, we were unable to replicate the results in the paper, particularly for the FID metric, where we observed a significant difference. Could you provide any suggestions?
The data of ATISS and DiffuScene are provided by the Midiffusion paper.
Pretrained is the weight you provide in the github.
Train by us is the weight we trained.
Thank you for your help.
The text was updated successfully, but these errors were encountered: