NN_compression icon indicating copy to clipboard operation
NN_compression copied to clipboard

I use GZIP to compress all of 10-50 markovity datasets but there is no different size among them

Open viviancui59 opened this issue 6 years ago • 4 comments

Should I try higher markovity dataset to compress using Gzip? The datasets are all generate by your generating code.

viviancui59 avatar May 23 '18 07:05 viviancui59

I try to use myself code to compress these dataset, but those compressed datasets' sizes are the same with each other, so I try to use GZIP to compress, the results are identical.

viviancui59 avatar May 23 '18 07:05 viviancui59

How did you use gzip to compress files? Did you compress directly on the generated .txt text files? I am wondering how much the encoding schema will affect the final comparision with NN_compression.

sonack avatar Jul 11 '18 02:07 sonack

How did you use gzip to compress files? Did you compress directly on the generated .txt text files? I am wondering how much the encoding schema will affect the final comparision with NN_compression.

How to compute compress ratio by loss fuction?师兄

AnsonHooL avatar Jul 05 '19 02:07 AnsonHooL

Based on my experiments, gzip should be able to compress markovity 10 dataset much better than markovity 50.

Can you post the command you used for generate the files (with the options).

@AnsonHooL: The loss value is compression ratio on the batch (without accounting for the 2/N factor for arithmetic coding)

kedartatwawadi avatar Jul 05 '19 02:07 kedartatwawadi