Replies: 1 comment 2 replies
-
Sorry I'm getting to this late. Yeah, the fast Herbie feature is an immature feature that needs more development and testing (thanks for testing it 😁). Depending on the amount of data you are requesting, I've noticed its faster to just download the full file and subset them after. I'll have to think about this issue. Here's some of my thought process... # This is the new way to use FastHerbie
from herbie import FastHerbie
# Create multiple Herbie objects
FH = FastHerbie(
DATES=["2022-11-29"],
model="ecmwf",
product="enfo",
fxx=range(3, 12, 3),
)
# You can look at those objects with this
FH.objects
# Download those files (subsets)
# This took almost 2 minutes
FH.download(searchString=":tp:sfc:")
# Open the data with Xarray
ds = FH.xarray(searchString=":tp:sfc:") But reading the data into xarray causes a error. I'll have to track down this bug for this case.
|
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I was trying to use Herbie to easily download and process the european ensemble data. Not sure if I don't understand Herbie-fast or what, because Herbie fast appears to download the entire dataset, and seems like it takes way longer to do one timestep than my loop for 6 days. I have the sample of code I am using below. The fast herbie that is commented out took an hour for maybe one time step? Not sure, when it started another loop, I killed it, because my other loop takes 45 minutes, however, I have not been able to download a complete dataset in the three days I have been trying, for any cycle.
The most common problem is one or more timesteps will have the number (member/pertubation) coordinates as 0 instead of a an array of length 50. If I go back an assign those 50, it's clear something weird happened as revealed by this spot check of a single point (look at the 05-03Z column):
I have not yet attempted to just download those times separately, but I would think this has to be a problem with the processing vs data, right? This is probably way more data than with a typical use-case, but I like me my ensemble data.
Other than these issues, kudos on this though, it makes dealing with this big data so so so so so so so much easier, and really helps elevate some serious science game.
Beta Was this translation helpful? Give feedback.
All reactions