hivemind
hivemind copied to clipboard
Decentralized deep learning in PyTorch. Built to train models on thousands of volunteers across the world.
This PR integrates blockwise quantization from [bitsandbytes](https://github.com/facebookresearch/bitsandbytes) as a new compression mechanism of Hivemind. The important part is that it is an *optional* compression protocol: the user should only install...
## The nature of this issue During #470 review there was a list of thing that were not crucial for the PR but ideally they should be done. Find problems...
Current plan: - working with https://stable-baselines3.readthedocs.io - trying to make a minimalistic example that uses hivemind.Optimizer TODO: - [x] make a PPO run with more than 1 peer - [x]...
Roadmap
This is a global project roadmap that states our priorities for the nearest future. These priorities can and should be disputed here or elsewhere, after which we will update the...
**Status:** This PR is an early draft intended to validate the design of `hivemind.DDPOptimizer`. I didn't run the code even once yet. **Co-authored-by:** @justheuristic
This is a collection of miscellaneous small updates that would make examples/albert more efficient or easier to understand. __Note 1:__ if you're looking for a more advanced example where many...
__problem:__ if many peers join at once, they will all pick one averager (latest at the time) as a target for loading initial state. This is causes choke points as...
We're using this dependency in one spot, where it can be replaced with ~5 lines of native code. Would be great to remove it
It's something we played with a few times but did not end up merging to master. I'm creating this issue so we wouldn't forget it. It would be great if...
- [ ] (important) Reducers should work concurrently - [ ] (important) Test with .isfinite(x), test that no other tensor values may corrupt CClip - [ ] Rename to `MeanReducer`,...