LLM-VM
LLM-VM copied to clipboard
irresponsible innovation. Try now at https://chat.dev/
Definition of done: Implement training large models using FSDP to accelerate training on large datasets. Reference: https://pytorch.org/blog/introducing-pytorch-fully-sharded-data-parallel-api/
Could possibly be done using `Scalene`
Enable large models that can't fit on the GPU to run inference by passing params back and forth between RAM and GPU-RAM
Currently LLM-VM does not support multiple GPU setups. Using runpod, I rented a setup with 2 RTX 3090 GPUs. Well running the local Bloom model example from the [docs](https://anarchy.ai/get_started/quickstart/completions). I...
We discussed YouTube videos in our latest triage but before we publish any videos, it would be good to have the video content as blogs.
open bounty for demo applications that work to add to a curated example gallery