GraphCNN and CPU memory usage #413
-
|
Hi AMPL staff. I am currently using a GraphCNN and CPU usage is massive. @2.5% of the final intended training data, CPU usage has approached 150GB. Questions:
Thanks. |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 4 replies
-
|
Hello,
|
Beta Was this translation helpful? Give feedback.
-
|
Thanks Stewart. Any idea what is the likely outcome and status on the DiskDataset? Also I've tried to address this problem using ECFP4s instead of graphs, and while I saw a 10% decrease in memory, and a 4x training speed, I noticed that training failed after 17 epochs, signified by a drop in r2 vs epoch. I suspect this is because ECFP4s are binary vectors and explosion problems occur, which graphs smoothing out can avoid (unless you have other insights). AMPL isn't really geared to troubleshoot this within its framework is it? If so, it seems AMPL is really not something that can train in the 100M, let alone 1B range is it? |
Beta Was this translation helpful? Give feedback.
There aren't any plans to implement the DiskDataset. I don't think it will be able to scale to the 100M or 1B range.
You can try experimenting with batch size, network size, weight decay, or learning rate. I don't think we have any tools for you to directly inspect the weights or deltas.