Daniel

Results 101 comments of Daniel

可以贡献下tts 用的是哪 个模型?可以转onnx或ncnn?

yes, you do.

> In the paper, it was mentioned that 48G of graphics memory can train 65B of LLaMA > > ``` > We present QLORA, an efficient finetuning approach that reduces...

Is there an example for training on multiple-GPUs?

ddp_find_unused_parameters False

load another model: ValueError: Target modules ['dense_h_to_4h', 'dense', 'dense_4h_to_h', 'query_key_value'] not found in the base model. Please check the target modules and try again. python qlora.py –learning_rate 0.0001 --model_name_or_path huggyllama/llama-7b

Yes. I tested it with 8* A100, it occupates 12G on each GPU and still not fast, especially alter the input length more than 500.

> @internlm-team You told us number of parameters at least :) We present InternLM, a multilingual foundational language model with 104B parameters. InternLM is pre-trained on a large corpora with...