Frank McSherry

Results 103 comments of Frank McSherry

Excellent, thanks! I'll try this out and see what the delta is. To be honest, I'm not 100% sure if the goal of the repo is to max out each...

It seems to have improved the times! I got a 10s build still, but a 294.73s second run. This is still a fair bit slower than differential, probably for the...

Can do. The `doop` and `galen` ones have clear output sizes. The `crdt` one is more awkward in that it only produces the "winning edit locations" not the characters that...

I think this is at least partially addressed with https://github.com/frankmcsherry/dynamic-datalog/commit/17a8b1f83451014ff4505d0685ed338808b54147. Let me know what you think.

More detail: * CRDT: Martin, Michael, and Mistral went over the outputs with me in person in Cambridge, and we validated that it was producing the right results on smaller...

On the road at the moment, but will investigate. The project is a bit old, so perhaps there is a lingering bug that should be fixed. The [`graph-map`](https://github.com/frankmcsherry/graph-map) project is...

I think this code has only been used on up to 4 billion edges (with the UK-2007) graph, but nothing prevents a larger graph. The main requirement is available memory...

And, if it helps, you can read more about performance on 1.6 billion edge and 3.7 billion edge graphs at: https://github.com/frankmcsherry/blog/blob/master/posts/2015-07-08.md https://github.com/frankmcsherry/blog/blob/master/posts/2015-07-31.md The short answer is, twenty iterations on the...

Sounds good; the 128B common crawl graph has identifiers that still fit in a u32, if that helps, but if you have your own larger graph you may need to...

Oh wow; how did I miss this issue from 2016? The pagerank values are stored in the `src` vector as the computation iterates. If you'd like them written out, the...