llama.cpp icon indicating copy to clipboard operation
llama.cpp copied to clipboard

Eval bug: llama_model_load: error loading model: error loading model hyperparameters: key not found in model: llama.context_length

Open www222fff opened this issue 9 months ago • 0 comments

Name and Version

python3 convert_lora_to_gguf.py /models/CodeLlama-13b-finetuned --base /models/CodeLlama-13b-Instruct-hf --outfile /models/CodeLlama-13b.gguf --outtype f32

INFO:gguf.gguf_writer:Writing the following files: INFO:gguf.gguf_writer:/models/CodeLlama-13b.gguf: n_tensors = 320, total_size = 104.9M Writing: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 105M/105M [00:00<00:00, 210Mbyte/s] INFO:lora-to-gguf:Model successfully exported to /models/CodeLlama-13b.gguf

root@7d293a7fbb1c:/app/llama.cpp# ./build/bin/llama-cli -m /models/CodeLlama-13b.gguf build: 5092 (d3bd7193) with cc (Debian 12.2.0-14) 12.2.0 for x86_64-linux-gnu main: llama backend init main: load the model and apply lora adapter, if any llama_model_loader: loaded meta data with 12 key-value pairs and 320 tensors from /models/CodeLlama-13b.gguf (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = llama llama_model_loader: - kv 1: general.type str = adapter llama_model_loader: - kv 2: adapter.type str = lora llama_model_loader: - kv 3: general.name str = CodeLlama 13b Finetuned llama_model_loader: - kv 4: general.finetune str = finetuned llama_model_loader: - kv 5: general.basename str = CodeLlama llama_model_loader: - kv 6: general.size_label str = 13B llama_model_loader: - kv 7: general.base_model.count u32 = 1 llama_model_loader: - kv 8: general.base_model.0.name str = models/CodeLlama 13b Instruct Hf llama_model_loader: - kv 9: general.base_model.0.repo_url str = https://huggingface.co//models/CodeLl... llama_model_loader: - kv 10: adapter.lora.alpha f32 = 32.000000 llama_model_loader: - kv 11: general.quantization_version u32 = 2 llama_model_loader: - type f32: 320 tensors print_info: file format = GGUF V3 (latest) print_info: file type = all F32 (guessed) print_info: file size = 100.00 MiB (32.00 BPW) llama_model_load: error loading model: error loading model hyperparameters: key not found in model: llama.context_length llama_model_load_from_file_impl: failed to load model common_init_from_params: failed to load model '/models/CodeLlama-13b.gguf' main: error: unable to load model

Operating systems

Linux

GGML backends

CPU

Hardware

cpu node cloud

Models

CodeLlama-13b-Instruct-hf

Problem description & steps to reproduce

What is the issue here?

First Bad Commit

No response

Relevant log output

NA

www222fff avatar Apr 10 '25 00:04 www222fff