Zygo
Zygo
Why CRC64? It was a familiar algorithm that produced 64 bits of output. Not the fastest, but perf says that other parts of bees take significantly more CPU time, so...
Now that we have an option parser, there should really be an option for this. Using the csum tree has a severe negative impact on the ability to find several...
csums are per on-disk block (which is not the same thing as a block read through the POSIX filesystem interface). The items in the csum tree are not extent-aligned, i.e....
cityhash > xxhash > crc64, saving both processing time and disk space. 2xCRC32C (like the function dduper uses) is still the decisive winner for hashing speed because that function is...
No objections at first glance. I'll have a deeper look later.
It's possible to check all the crawlers to see if they're all idle (deferred), and exit if that is true. The complicating factor is "all the crawlers": * during startup...
Walltime doesn't need bees to implement: ``` perl -e 'alarm(8*3600); exec @ARGV' bees --one-shot /your/fs ``` (well, that would work if bees handled `SIGALRM` the same way as `SIGTERM`) We...
OK, I accept that rationale for `--oneshot` / `-1`. My point was that you could just set up a cron job or external timer to make sure bees doesn't take...
The perl man page talks about `sleep` on general Unix platforms, not `sleep` on Linux. On old crappy Unix, `sleep` is a library function which sets up a `SIGALRM` with...
OK, to summarize the feature request: * add `--oneshot` (`-1`) option to run bees until all crawlers are past the current (measured at bees startup time) filesystem transid, then exit...