Plans for Different Sized Models - 1.5B or 3B
Thank you for this amazing work!
Is there a plan to train and release smaller (1.5B to 3B) or larger (14B to 70B) sized models for LLaDA?
Thank you for your interest! Currently, we don't have a specific plan or timeline.
Thank you for your interest! Currently, we don't have a specific plan or timeline.
Would you like to ask if it is applicable to general model quantization techniques? llada seems to perform well on coding tasks, and if it can reduce the burden of reasoning, it is believed to lead to widespread use:D
Thank you for your attention!
Currently, we are just a very small team and don't have the capacity to carry out the model quantification work at the moment. I've noticed that some people in the community are attempting to do this. Perhaps you could refer to their work.
Is it public who's working on this? And if so, do they have a public page or something?