What benchmark tasks shall we add to mofdscribe
?
#244
Replies: 3 comments 5 replies
-
Another thing I have been thinking a lot about is how a "blind test" could look like. That is, a challenge where the test set is not public (and we hence can avoid adaptive overfitting or other issues). I feel it would be great if this would not come from one of the main academic MOF groups (as they would all want to participate) but is organized by some external entity (CSD? @ml-evs or industry) |
Beta Was this translation helpful? Give feedback.
-
@FMcil do you think we can add anything interesting from PrIsMa? I think we do not have that many data points yet (<2000?). Do we have some other process-model outputs? Would also nice to expose the RSM materials as dataset, but I do not know what labels we have. |
Beta Was this translation helpful? Give feedback.
-
@arosen93 do you have any wishes for what benchmark tasks we should include? |
Beta Was this translation helpful? Give feedback.
-
An important aspect we want to promote with
mofdscribe
is to make it easier to compare models developed for reticular chemistry (common task framework, see Donoho). Even though there are clearly issues with benchmarks (the community tends to overfit them) we still think that they would be very useful.Currently, we have some benchmark tasks implemented in
mofdscribe
but we would like to revise and extend this. For this we would welcome any feedback.In the current concept of benchmark tasks, each benchmark describes:
Cory Simon (@SimonEnsemble) created an overview of existing datasets: https://github.com/SimonEnsemble/porous-material-AI-gym
Beta Was this translation helpful? Give feedback.
All reactions