llama
llama copied to clipboard
GPU hardware for inference
Can you please specify/estimate the hardware required to run each model? It will help me choose the correct EC2 instance. Many thanks!
for inference, it's like a 26GB VRAM card. for training, when i use 48GB, unfortunately, OOM.... still trying~ if you find proper devices, please tell me.
@kriskrisliu According to what I read, training 7B model requires at least 4 x 80GB GPUs.