open_llama icon indicating copy to clipboard operation
open_llama copied to clipboard

OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset

Results 55 open_llama issues
Sort by recently updated
recently updated
newest added
trafficstars

Does this still a bug for tokenization? I want to use this for code. Thanks!

I use FastChat as the framework for both training and dialog-based inference, and FastChat supports Meta/Llama. I was excited to try the 3B state Open-Llama model, and the FastChat finetuning...

Where is the code to train llama on tpu's?

My code ``` ef reinitialize_weights_gpt_neox_20B_inspired_4_llama2(model): """ Note: we nearly gpt-neox_20B (2022) & llama1 , llama2 (2019) does not say how they init I think gpt-neox_20B & llama2 both have pre-layernorm,...

I have a question regarding training stability. I downloaded the complete dataset of Redpajama v1 from Hugging Face and followed the parameter settings from the Llama1 paper for data mixture...

I am quite new to this space, are there any resources anyone would recommend? Further, what I am looking to do is have a set of "Context" that I would...

Use Llama2 model and train on all latest and more efficient (like SlimPajama vs redpajama) open datasets ? Just for the base model, then maybe open-assistant team can rlhf it

How can you call yourself open anything when you can't even release your code?