YaLM-100B
YaLM-100B copied to clipboard
Run on networked nodes
Thanks for open-sourcing this! Because the GPU ram requirements are so high, it's hard to rent a large enough single node from any of the major cloud providers. How can you run it in inference mode networked between multiple physical machines?
Thanks!