Peter Dudfield
Peter Dudfield
Perhaps 'seeding' them is a good comprise?
The validation script is 'abit' / 'a lot' out of data. Itll need some work to update. The good thing is the Batch validates each data source as we go....
I remember using something like 'gzip' made the files smaller. But then it took longer to load. http://xarray.pydata.org/en/stable/generated/xarray.Dataset.to_netcdf.html I'm not sure on the right balance here also just some general...
Might as well - small amount of work, for small win
Moving back here > > > Do you mind expanding on what needs doing for the NWP data? > > > > > > Sure! > > The short answer...
Would be interesting to know if some default compression would be the same as doing this rescaling (see above). Probably only way it to try it out ...
I was just about to type the same, so low hanging fruit, change it to float16 the test data was float32 = 13M float16 = 4.6M int16 = 3 M...
Ive pushed that now to PR #335
Ill keep it open, incase we do want to go down the int8/int16 route
I think needs a rewrite of the validation script too Potentially something we can live we out as the 'manager.py' now does some validation