Francesc Alted
Francesc Alted
This is a good point. I think we are not doing the correct thing with `blosc2.compress`, but `blosc2.compress2` should be fine. I still need to think a bit more about...
Perhaps it would be better to start with [building](https://docs.microsoft.com/en-us/azure/devops/pipelines/artifacts/pipeline-artifacts?view=azure-devops&tabs=yaml) a [Azure pipeline artifact](https://docs.microsoft.com/en-us/azure/devops/pipelines/artifacts/artifacts-overview?view=azure-devops) and [release it in PyPI](https://docs.microsoft.com/en-us/azure/devops/pipelines/artifacts/pypi?view=azure-devops&tabs=yaml).
Closing due to inactivity.
With just 22400 items, your chunk seems too small for blosc to split it in different blocks (the only way to leverage multithreading). Try with something that is at least...
This issue stems from https://github.com/Blosc/c-blosc2/issues/467
That would be great. The new API for C-Blosc2 is backward compatible with C-Blosc, so this should be easy. Just remember that the C-Blosc2 binary format is backward compatible, but...
That's correct.
I don't think so. At first I was trying to keep a format that was forward-compatible, but it was too much hassle, and decided not to do it.
We would need to try, but I am pretty sure that an error will be raised when using C-Blosc1 with chunks generated with C-Blosc2. On the other hand, the [way...
After pondering a bit more about Blosc/Blosc2 compatibility, I think a better approach is to make the two filters totally separate. So the current hdf5-blosc will continue supporting just the...