snappy-java
snappy-java copied to clipboard
Snappy compressor/decompressor for Java
There should be an assertion so that length should not exceed the int size limit.
https://github.com/kiyo-masui/bitshuffle 0.5.1 is the latest version, but due to API incompatibilities and various build failures when using cross compilers, it's been difficult to change the bitshuffle version to use. The...
The underlying BitShuffle library works internally with an algorithm that was designed for the little-endian format. For this reason, the input data must always be passed in little-endian format. However,...
For data encoding, sometimes integers with intermediate sizes (e.g. of 3 or 5 bytes) are used (e.g. to reduce bandwith). These arrays currently cannot be passed to the external BitShuffle...
Full details here: https://stackoverflow.com/questions/51069767/maven-unknown-packaging-bundle-error-from-a-dependency-packaging-as-bundle
Currently (v1.1.8.4) snappy-java fails to use non-direct ByteBuffers with the following exception thrown: ``` org.xerial.snappy.SnappyError: [NOT_A_DIRECT_BUFFER] input is not a direct buffer at org.xerial.snappy.Snappy.compress(Snappy.java:141) ```
When testing upgrade to spark 3.1.1 I've noticed the compression of repeated INT64 columns compression got worse. https://stackoverflow.com/questions/67413589/parquet-compression-degradation-when-upgrading-spark/67455721#67455721 Reading [this file](https://drive.google.com/file/d/1FZx_qAmoX1HDpAVplvFnl2iC83siOCTE/view?usp=sharing) saved with snappy 1.1.2.6, and writing it with higher...
I had this error while I tried to decompress a large file. Size of sample.snappy is 10 GB. Is there a way around it. Code source : [Source](https://partners-intl.aliyun.com/help/doc-detail/108942.htm) ``` String...
SnappyOutputStream uses a buffer management system to reduce memory pressure and GC overhead. The same mechanism would be helpful in SnappyInputStream as well. We want to use Snappy for protocol...