Neo LLama

Results 5 comments of Neo LLama

> 全量finetune,ZeRO3,设置output_router_logits=True。训练过程中会突然卡住,GPU利用率突然到100% ![image](https://private-user-images.githubusercontent.com/96909430/321122435-096c34cf-fb9c-4e1e-b694-47a5a104d6b9.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MTMzNjA5MTAsIm5iZiI6MTcxMzM2MDYxMCwicGF0aCI6Ii85NjkwOTQzMC8zMjExMjI0MzUtMDk2YzM0Y2YtZmI5Yy00ZTFlLWI2OTQtNDdhNWExMDRkNmI5LnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNDA0MTclMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjQwNDE3VDEzMzAxMFomWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPWZjYTYxZjk4ZTkyODJjNzRkZGI2ZjAwMDE3ZmE1YWI1OWRmMmY4MDEyNzQ0ZGYwNmM2OWY4MDlmOWY5NGUxOWEmWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0JmFjdG9yX2lkPTAma2V5X2lkPTAmcmVwb19pZD0wIn0.rV3YPjOhzllPlrg94KOjiEjUaHpOKJqx_653NeWWc54) 请问有解决吗

> hello! I also met some problems when running the msra2mrc.py?do you know how to deal with the problem of invilid inputs?thank you so much if you could give me...

> ```python > torch.set_default_tensor_type(torch.cuda.HalfTensor) > ``` i have same problem when i train llama3, in modeling_llama.py 1095: ``` causal_mask = torch.triu(causal_mask, diagonal=1) ``` i fix this by : ``` causal_mask...

excuse me ,can you help me convert the raw videos and text to TFRecords,i have downloaded the two datasets ![image](https://github.com/google-research/google-research/assets/55910045/6b5dccfc-7f01-4198-88b2-f34379af8478) how can i convert them? thank you

> @cooper12121, please share repro steps, such as scripts, full stack trace, and ds_config. **I use transformers' Trainer to train my model, Here is the full error report:** ``` 11.219.19.47:...