Hao Zhang
Hao Zhang
> I have addressed the comments. Do you have any further suggestions? Will review tomorrow and see whether more changes are needed.
@wjykl22 Thanks for your interest! We're currently working on cleaning up the dataset and will make it available in the Neurips week (i.e., next week). For the AutoSync code, the...
@ZeyaWang : are you working on this one now?
@Ezra-H CI still fails, please check & fix
@rjiang-ptm thanks! @merrymercy @zhuohan123 please comment
Is TPU+pipeshard still in our scope? @merrymercy @ZYHowell
See this paper by Alpa team: https://arxiv.org/pdf/2211.05322.pdf Related PR: #773
@sachit-menon @xhluca Hi, We recently integrate OPT-175B with the Berkeley-backed system [Alpa](https://github.com/alpa-projects/alpa). I guess you can try Alpa which exactly allows you to use more heterogeneous GPU setups to train/serve...
We have done some study and our temporal conclusion is that the backbone model Cerebras-GPT has much lower quality than the llama, due to pertaining dataset size, and data filtering...