Speeding up a model with a lot of data #125
-
Question via email:
|
Beta Was this translation helpful? Give feedback.
Replies: 1 comment
-
The biggest thing to speed things up is to use a coarser mesh. Also, some random field structures will be faster than others. Adding spatiotemporal fields will be much slower than just spatial fields. Depending what you’re doing, sometimes A large number of With > 500k observations, you could also consider gridding your data and changing the family accordingly. E.g. instead of binomial, grid the data, use the centroids, and model the count per cell as negative binomial. Or similar for biomass etc. Or downsampling for model experimentation. Yes, you could certainly compare models with coarser meshes and then increase the resolution. You can often get away with surprisingly coarse meshes. At some point the models become overfit if the mesh becomes too fine, which isn’t always appreciated, and is only seen through cross validation (we're working on a paper looking at this). You can also try setting up parallel processing: e.g. But check that’s actually doing anything and actually faster. I’m not sure if that’s working on Windows. At some point more cores can definitely be slower. |
Beta Was this translation helpful? Give feedback.
silent = FALSE
just gives you progress info during fitting, which can be helpful to monitor progress on big models, but won't speed things up.The biggest thing to speed things up is to use a coarser mesh. Also, some random field structures will be faster than others. Adding spatiotemporal fields will be much slower than just spatial fields. Depending what you’re doing, sometimes
reml = TRUE
is faster, but I wouldn’t use that for index standardization, if that’s what you’re doing.A large number of
extra_time
slices will slow things down.With > 500k observations, you could also consider gridding your data and changing the family accordingly. E.g. instead of binomial, grid the data, use t…