Skip to content

Suitability for large datasets #22

Answered by TheoMichelot
roaldarbol asked this question in Q&A
Discussion options

You must be logged in to vote

Hi, here are a few thoughts about using hmmTMB on large data sets.

  1. We wrote the package with speed in mind, e.g., the likelihood function is implemented in C++ using TMB. That being said, we also wanted hmmTMB to be flexible (e.g., many observation distributions, general covariate dependence), so we had to sacrifice speed in some places.

  2. I just ran a simulation to fit a 2-state Gaussian HMM to 2.5 million data points, and it took around 3 minutes on my laptop. Here I was fitting the true model, so it is likely that it would take somewhat longer in a real data analysis, but it should be feasible.

  3. Computing time will very strongly depend on the model formulation, and model fitting can…

Replies: 2 comments 12 replies

Comment options

You must be logged in to vote
9 replies
@roaldarbol
Comment options

@roaldarbol
Comment options

@TheoMichelot
Comment options

@roaldarbol
Comment options

@TheoMichelot
Comment options

Answer selected by TheoMichelot
Comment options

You must be logged in to vote
3 replies
@roaldarbol
Comment options

@TheoMichelot
Comment options

@roaldarbol
Comment options

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants