petals icon indicating copy to clipboard operation
petals copied to clipboard

🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading

Results 92 petals issues
Sort by recently updated
recently updated
newest added

Really cool project! Feels like it has the potential to be a game-changer for poor students like myself 😮 Does anyone know how/where to view the following statistics of the...

Having to specifically hard code IP adresses makes it very hard to run petals on a SLURM cluster. There I submit batch jobs that are then run on some node...

The primary motivation, is - to get as much high bandwidth memory, in a low cost way (thanks to its unified memory model) - to be easily used for training...

Hello folks, Trying to run a private swarm on a 7x Volta-generation GPUs. As suggested by docs, i've set torch_dtype to float16 and NUM_BLOCKS to 10 (these are 32GB GPUs)...

* task_pool * [x] remove interate_minibatches, once https://github.com/learning-at-home/hivemind/pull/506 is merged * [ ] batch sequences of similar length * runtime * [x] do not log "0 parameters" on init (misleading)...

__Current tasks:__ - [ ] prototype bloom points system @borzunov (#6 ) - [x] local tensor parallelism ( #143 , using [BlackSamorez/tesnor_parallel](https://github.com/BlackSamorez/tensor_parallel) by @blacksamorez and @IaroslavLisniak ) - [x] increase...

[in consultation with @mryab] The max input token length is 2048 right now. It would be nice to process more than 2048 tokens through the distributed BLOOM. Increasing the max...

This PR is intended as a backup if we experience any problems with finetuning in the current branch. TL;DR previously we tested that finetuning works with compression as in this...