distributed-llm topic
List
distributed-llm repositories
distributed-llama
894
Stars
63
Forks
Watchers
Tensor parallelism is all you need. Run LLMs on weak devices or make powerful devices even more powerful by distributing the workload and dividing the RAM usage.