llama.cpp icon indicating copy to clipboard operation
llama.cpp copied to clipboard

Running convert fails with BadZipFile (Bad CRC-32)

Open itsJoKr opened this issue 1 year ago • 5 comments

Prerequisites

Please answer the following questions for yourself before submitting an issue.

  • [x] I am running the latest code. Development is very rapid so there are no tagged versions as of now.
  • [x] I carefully followed the README.md.
  • [x] I searched using keywords relevant to my issue to make sure that I am creating a new issue that is not already open (or closed).
  • [x] I reviewed the Discussions, and have a new bug or useful enhancement to share.

Expected Behavior

The convert runs successfully.

Current Behavior

The convert fails with error like

  File "/Library/Frameworks/Python.framework/Versions/3.12/lib/python3.12/zipfile/__init__.py", line 996, in _update_crc
    raise BadZipFile("Bad CRC-32 for file %r" % self.name)

Environment and Context

Please provide detailed information about your computer setup. This is important in case the issue is not reproducible except for under certain specific conditions.

MacBook Pro, M1 Pro, MacOS Sonoma Python: 3.12.0 Make: GNU Make 3.81

I tried to run for both OpenLlama-3B and Llama-7B-chat, same error.

Steps to Reproduce

Please provide detailed steps for reproducing the issue. We are not sitting in front of your screen, so the more detail the better.

  1. Clone llama and llama.cpp
  2. Make llama.cpp
  3. Download llama 7B-chat model
  4. Run convert

Failure Logs

llama.cpp$ git log | head 
commit 5a7d3125e7c24f223659b7f0b7aa7736986e92c0

llama.cpp$ python3 --version
Python 3.12.0

llama.cpp$ pip list | egrep "torch|numpy|sentencepiece"
numpy              1.26.2
sentencepiece      0.1.99

itsJoKr avatar Dec 07 '23 16:12 itsJoKr

same issue for all of my models, worked before git pull + recomp

skidd-level-100 avatar Dec 29 '23 22:12 skidd-level-100

This seems to have been fixed on the last commit.

LIMPIX31 avatar Jan 09 '24 09:01 LIMPIX31

This seems to have been fixed on the last commit.

not for me still 'convert.py path/to/model' and 'convert.py path/to/model.bin' fail with the last to lines being:

raise BadZipFile("Bad CRC-32 for file %r" % self.name)

zipfile.BadZipFile: Bad CRC-32 for file 'archive/data/13'

skidd-level-100 avatar Jan 11 '24 08:01 skidd-level-100

I've got the same issue. Here's my test run after downloading the model from hf:

python3.12 llama.cpp/convert.py aiopsmodel-hfv2 --outfile llama-2-7b-aiopsfinetunedv2-q8_0-gguf --outtype q8_0 Loading model file aiopsmodel-hfv2/pytorch_model-00001-of-00002.bin Loading model file aiopsmodel-hfv2/pytorch_model-00001-of-00002.bin Loading model file aiopsmodel-hfv2/pytorch_model-00002-of-00002.bin params = Params(n_vocab=32001, n_embd=4096, n_layer=32, n_ctx=4096, n_ff=11008, n_head=32, n_head_kv=32, n_experts=None, n_experts_used=None, f_norm_eps=1e-05, rope_scaling_type=None, f_rope_freq_base=None, f_rope_scale=None, n_orig_ctx=None, rope_finetuned=None, ftype=<GGMLFileType.MostlyQ8_0: 7>, path_model=PosixPath('aiopsmodel-hfv2')) Found vocab files: {'tokenizer.model': PosixPath('aiopsmodel-hfv2/tokenizer.model'), 'vocab.json': None, 'tokenizer.json': PosixPath('aiopsmodel-hfv2/tokenizer.json')} Loading vocab file 'aiopsmodel-hfv2/tokenizer.model', type 'spm' Vocab info: <SentencePieceVocab with 32000 base tokens and 1 added tokens> Special vocab info: <SpecialVocab with 0 merges, special tokens {'bos': 1, 'eos': 2, 'unk': 0, 'pad': 0}, add special tokens unset> Permuting layer 0 Permuting layer 1 Permuting layer 2 Permuting layer 3 Permuting layer 4 Permuting layer 5 Permuting layer 6 Permuting layer 7 Permuting layer 8 Permuting layer 9 Permuting layer 10 Permuting layer 11 Permuting layer 12 Permuting layer 13 Permuting layer 14 Permuting layer 15 Permuting layer 16 Permuting layer 17 Permuting layer 18 Permuting layer 19 Permuting layer 20 Permuting layer 21 Permuting layer 22 Permuting layer 23 Permuting layer 24 Permuting layer 25 Permuting layer 26 Permuting layer 27 Permuting layer 28 Permuting layer 29 Permuting layer 30 Permuting layer 31 model.embed_tokens.weight -> token_embd.weight | F16 | [32000, 4096] model.layers.0.self_attn.q_proj.weight -> blk.0.attn_q.weight | F16 | [4096, 4096] model.layers.0.self_attn.k_proj.weight -> blk.0.attn_k.weight | F16 | [4096, 4096] model.layers.0.self_attn.v_proj.weight -> blk.0.attn_v.weight | F16 | [4096, 4096] model.layers.0.self_attn.o_proj.weight -> blk.0.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.0.attn_rot_embd model.layers.0.mlp.gate_proj.weight -> blk.0.ffn_gate.weight | F16 | [11008, 4096] model.layers.0.mlp.up_proj.weight -> blk.0.ffn_up.weight | F16 | [11008, 4096] model.layers.0.mlp.down_proj.weight -> blk.0.ffn_down.weight | F16 | [4096, 11008] model.layers.0.input_layernorm.weight -> blk.0.attn_norm.weight | F16 | [4096] model.layers.0.post_attention_layernorm.weight -> blk.0.ffn_norm.weight | F16 | [4096] model.layers.1.self_attn.q_proj.weight -> blk.1.attn_q.weight | F16 | [4096, 4096] model.layers.1.self_attn.k_proj.weight -> blk.1.attn_k.weight | F16 | [4096, 4096] model.layers.1.self_attn.v_proj.weight -> blk.1.attn_v.weight | F16 | [4096, 4096] model.layers.1.self_attn.o_proj.weight -> blk.1.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.1.attn_rot_embd model.layers.1.mlp.gate_proj.weight -> blk.1.ffn_gate.weight | F16 | [11008, 4096] model.layers.1.mlp.up_proj.weight -> blk.1.ffn_up.weight | F16 | [11008, 4096] model.layers.1.mlp.down_proj.weight -> blk.1.ffn_down.weight | F16 | [4096, 11008] model.layers.1.input_layernorm.weight -> blk.1.attn_norm.weight | F16 | [4096] model.layers.1.post_attention_layernorm.weight -> blk.1.ffn_norm.weight | F16 | [4096] model.layers.2.self_attn.q_proj.weight -> blk.2.attn_q.weight | F16 | [4096, 4096] model.layers.2.self_attn.k_proj.weight -> blk.2.attn_k.weight | F16 | [4096, 4096] model.layers.2.self_attn.v_proj.weight -> blk.2.attn_v.weight | F16 | [4096, 4096] model.layers.2.self_attn.o_proj.weight -> blk.2.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.2.attn_rot_embd model.layers.2.mlp.gate_proj.weight -> blk.2.ffn_gate.weight | F16 | [11008, 4096] model.layers.2.mlp.up_proj.weight -> blk.2.ffn_up.weight | F16 | [11008, 4096] model.layers.2.mlp.down_proj.weight -> blk.2.ffn_down.weight | F16 | [4096, 11008] model.layers.2.input_layernorm.weight -> blk.2.attn_norm.weight | F16 | [4096] model.layers.2.post_attention_layernorm.weight -> blk.2.ffn_norm.weight | F16 | [4096] model.layers.3.self_attn.q_proj.weight -> blk.3.attn_q.weight | F16 | [4096, 4096] model.layers.3.self_attn.k_proj.weight -> blk.3.attn_k.weight | F16 | [4096, 4096] model.layers.3.self_attn.v_proj.weight -> blk.3.attn_v.weight | F16 | [4096, 4096] model.layers.3.self_attn.o_proj.weight -> blk.3.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.3.attn_rot_embd model.layers.3.mlp.gate_proj.weight -> blk.3.ffn_gate.weight | F16 | [11008, 4096] model.layers.3.mlp.up_proj.weight -> blk.3.ffn_up.weight | F16 | [11008, 4096] model.layers.3.mlp.down_proj.weight -> blk.3.ffn_down.weight | F16 | [4096, 11008] model.layers.3.input_layernorm.weight -> blk.3.attn_norm.weight | F16 | [4096] model.layers.3.post_attention_layernorm.weight -> blk.3.ffn_norm.weight | F16 | [4096] model.layers.4.self_attn.q_proj.weight -> blk.4.attn_q.weight | F16 | [4096, 4096] model.layers.4.self_attn.k_proj.weight -> blk.4.attn_k.weight | F16 | [4096, 4096] model.layers.4.self_attn.v_proj.weight -> blk.4.attn_v.weight | F16 | [4096, 4096] model.layers.4.self_attn.o_proj.weight -> blk.4.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.4.attn_rot_embd model.layers.4.mlp.gate_proj.weight -> blk.4.ffn_gate.weight | F16 | [11008, 4096] model.layers.4.mlp.up_proj.weight -> blk.4.ffn_up.weight | F16 | [11008, 4096] model.layers.4.mlp.down_proj.weight -> blk.4.ffn_down.weight | F16 | [4096, 11008] model.layers.4.input_layernorm.weight -> blk.4.attn_norm.weight | F16 | [4096] model.layers.4.post_attention_layernorm.weight -> blk.4.ffn_norm.weight | F16 | [4096] model.layers.5.self_attn.q_proj.weight -> blk.5.attn_q.weight | F16 | [4096, 4096] model.layers.5.self_attn.k_proj.weight -> blk.5.attn_k.weight | F16 | [4096, 4096] model.layers.5.self_attn.v_proj.weight -> blk.5.attn_v.weight | F16 | [4096, 4096] model.layers.5.self_attn.o_proj.weight -> blk.5.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.5.attn_rot_embd model.layers.5.mlp.gate_proj.weight -> blk.5.ffn_gate.weight | F16 | [11008, 4096] model.layers.5.mlp.up_proj.weight -> blk.5.ffn_up.weight | F16 | [11008, 4096] model.layers.5.mlp.down_proj.weight -> blk.5.ffn_down.weight | F16 | [4096, 11008] model.layers.5.input_layernorm.weight -> blk.5.attn_norm.weight | F16 | [4096] model.layers.5.post_attention_layernorm.weight -> blk.5.ffn_norm.weight | F16 | [4096] model.layers.6.self_attn.q_proj.weight -> blk.6.attn_q.weight | F16 | [4096, 4096] model.layers.6.self_attn.k_proj.weight -> blk.6.attn_k.weight | F16 | [4096, 4096] model.layers.6.self_attn.v_proj.weight -> blk.6.attn_v.weight | F16 | [4096, 4096] model.layers.6.self_attn.o_proj.weight -> blk.6.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.6.attn_rot_embd model.layers.6.mlp.gate_proj.weight -> blk.6.ffn_gate.weight | F16 | [11008, 4096] model.layers.6.mlp.up_proj.weight -> blk.6.ffn_up.weight | F16 | [11008, 4096] model.layers.6.mlp.down_proj.weight -> blk.6.ffn_down.weight | F16 | [4096, 11008] model.layers.6.input_layernorm.weight -> blk.6.attn_norm.weight | F16 | [4096] model.layers.6.post_attention_layernorm.weight -> blk.6.ffn_norm.weight | F16 | [4096] model.layers.7.self_attn.q_proj.weight -> blk.7.attn_q.weight | F16 | [4096, 4096] model.layers.7.self_attn.k_proj.weight -> blk.7.attn_k.weight | F16 | [4096, 4096] model.layers.7.self_attn.v_proj.weight -> blk.7.attn_v.weight | F16 | [4096, 4096] model.layers.7.self_attn.o_proj.weight -> blk.7.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.7.attn_rot_embd model.layers.7.mlp.gate_proj.weight -> blk.7.ffn_gate.weight | F16 | [11008, 4096] model.layers.7.mlp.up_proj.weight -> blk.7.ffn_up.weight | F16 | [11008, 4096] model.layers.7.mlp.down_proj.weight -> blk.7.ffn_down.weight | F16 | [4096, 11008] model.layers.7.input_layernorm.weight -> blk.7.attn_norm.weight | F16 | [4096] model.layers.7.post_attention_layernorm.weight -> blk.7.ffn_norm.weight | F16 | [4096] model.layers.8.self_attn.q_proj.weight -> blk.8.attn_q.weight | F16 | [4096, 4096] model.layers.8.self_attn.k_proj.weight -> blk.8.attn_k.weight | F16 | [4096, 4096] model.layers.8.self_attn.v_proj.weight -> blk.8.attn_v.weight | F16 | [4096, 4096] model.layers.8.self_attn.o_proj.weight -> blk.8.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.8.attn_rot_embd model.layers.8.mlp.gate_proj.weight -> blk.8.ffn_gate.weight | F16 | [11008, 4096] model.layers.8.mlp.up_proj.weight -> blk.8.ffn_up.weight | F16 | [11008, 4096] model.layers.8.mlp.down_proj.weight -> blk.8.ffn_down.weight | F16 | [4096, 11008] model.layers.8.input_layernorm.weight -> blk.8.attn_norm.weight | F16 | [4096] model.layers.8.post_attention_layernorm.weight -> blk.8.ffn_norm.weight | F16 | [4096] model.layers.9.self_attn.q_proj.weight -> blk.9.attn_q.weight | F16 | [4096, 4096] model.layers.9.self_attn.k_proj.weight -> blk.9.attn_k.weight | F16 | [4096, 4096] model.layers.9.self_attn.v_proj.weight -> blk.9.attn_v.weight | F16 | [4096, 4096] model.layers.9.self_attn.o_proj.weight -> blk.9.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.9.attn_rot_embd model.layers.9.mlp.gate_proj.weight -> blk.9.ffn_gate.weight | F16 | [11008, 4096] model.layers.9.mlp.up_proj.weight -> blk.9.ffn_up.weight | F16 | [11008, 4096] model.layers.9.mlp.down_proj.weight -> blk.9.ffn_down.weight | F16 | [4096, 11008] model.layers.9.input_layernorm.weight -> blk.9.attn_norm.weight | F16 | [4096] model.layers.9.post_attention_layernorm.weight -> blk.9.ffn_norm.weight | F16 | [4096] model.layers.10.self_attn.q_proj.weight -> blk.10.attn_q.weight | F16 | [4096, 4096] model.layers.10.self_attn.k_proj.weight -> blk.10.attn_k.weight | F16 | [4096, 4096] model.layers.10.self_attn.v_proj.weight -> blk.10.attn_v.weight | F16 | [4096, 4096] model.layers.10.self_attn.o_proj.weight -> blk.10.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.10.attn_rot_embd model.layers.10.mlp.gate_proj.weight -> blk.10.ffn_gate.weight | F16 | [11008, 4096] model.layers.10.mlp.up_proj.weight -> blk.10.ffn_up.weight | F16 | [11008, 4096] model.layers.10.mlp.down_proj.weight -> blk.10.ffn_down.weight | F16 | [4096, 11008] model.layers.10.input_layernorm.weight -> blk.10.attn_norm.weight | F16 | [4096] model.layers.10.post_attention_layernorm.weight -> blk.10.ffn_norm.weight | F16 | [4096] model.layers.11.self_attn.q_proj.weight -> blk.11.attn_q.weight | F16 | [4096, 4096] model.layers.11.self_attn.k_proj.weight -> blk.11.attn_k.weight | F16 | [4096, 4096] model.layers.11.self_attn.v_proj.weight -> blk.11.attn_v.weight | F16 | [4096, 4096] model.layers.11.self_attn.o_proj.weight -> blk.11.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.11.attn_rot_embd model.layers.11.mlp.gate_proj.weight -> blk.11.ffn_gate.weight | F16 | [11008, 4096] model.layers.11.mlp.up_proj.weight -> blk.11.ffn_up.weight | F16 | [11008, 4096] model.layers.11.mlp.down_proj.weight -> blk.11.ffn_down.weight | F16 | [4096, 11008] model.layers.11.input_layernorm.weight -> blk.11.attn_norm.weight | F16 | [4096] model.layers.11.post_attention_layernorm.weight -> blk.11.ffn_norm.weight | F16 | [4096] model.layers.12.self_attn.q_proj.weight -> blk.12.attn_q.weight | F16 | [4096, 4096] model.layers.12.self_attn.k_proj.weight -> blk.12.attn_k.weight | F16 | [4096, 4096] model.layers.12.self_attn.v_proj.weight -> blk.12.attn_v.weight | F16 | [4096, 4096] model.layers.12.self_attn.o_proj.weight -> blk.12.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.12.attn_rot_embd model.layers.12.mlp.gate_proj.weight -> blk.12.ffn_gate.weight | F16 | [11008, 4096] model.layers.12.mlp.up_proj.weight -> blk.12.ffn_up.weight | F16 | [11008, 4096] model.layers.12.mlp.down_proj.weight -> blk.12.ffn_down.weight | F16 | [4096, 11008] model.layers.12.input_layernorm.weight -> blk.12.attn_norm.weight | F16 | [4096] model.layers.12.post_attention_layernorm.weight -> blk.12.ffn_norm.weight | F16 | [4096] model.layers.13.self_attn.q_proj.weight -> blk.13.attn_q.weight | F16 | [4096, 4096] model.layers.13.self_attn.k_proj.weight -> blk.13.attn_k.weight | F16 | [4096, 4096] model.layers.13.self_attn.v_proj.weight -> blk.13.attn_v.weight | F16 | [4096, 4096] model.layers.13.self_attn.o_proj.weight -> blk.13.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.13.attn_rot_embd model.layers.13.mlp.gate_proj.weight -> blk.13.ffn_gate.weight | F16 | [11008, 4096] model.layers.13.mlp.up_proj.weight -> blk.13.ffn_up.weight | F16 | [11008, 4096] model.layers.13.mlp.down_proj.weight -> blk.13.ffn_down.weight | F16 | [4096, 11008] model.layers.13.input_layernorm.weight -> blk.13.attn_norm.weight | F16 | [4096] model.layers.13.post_attention_layernorm.weight -> blk.13.ffn_norm.weight | F16 | [4096] model.layers.14.self_attn.q_proj.weight -> blk.14.attn_q.weight | F16 | [4096, 4096] model.layers.14.self_attn.k_proj.weight -> blk.14.attn_k.weight | F16 | [4096, 4096] model.layers.14.self_attn.v_proj.weight -> blk.14.attn_v.weight | F16 | [4096, 4096] model.layers.14.self_attn.o_proj.weight -> blk.14.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.14.attn_rot_embd model.layers.14.mlp.gate_proj.weight -> blk.14.ffn_gate.weight | F16 | [11008, 4096] model.layers.14.mlp.up_proj.weight -> blk.14.ffn_up.weight | F16 | [11008, 4096] model.layers.14.mlp.down_proj.weight -> blk.14.ffn_down.weight | F16 | [4096, 11008] model.layers.14.input_layernorm.weight -> blk.14.attn_norm.weight | F16 | [4096] model.layers.14.post_attention_layernorm.weight -> blk.14.ffn_norm.weight | F16 | [4096] model.layers.15.self_attn.q_proj.weight -> blk.15.attn_q.weight | F16 | [4096, 4096] model.layers.15.self_attn.k_proj.weight -> blk.15.attn_k.weight | F16 | [4096, 4096] model.layers.15.self_attn.v_proj.weight -> blk.15.attn_v.weight | F16 | [4096, 4096] model.layers.15.self_attn.o_proj.weight -> blk.15.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.15.attn_rot_embd model.layers.15.mlp.gate_proj.weight -> blk.15.ffn_gate.weight | F16 | [11008, 4096] model.layers.15.mlp.up_proj.weight -> blk.15.ffn_up.weight | F16 | [11008, 4096] model.layers.15.mlp.down_proj.weight -> blk.15.ffn_down.weight | F16 | [4096, 11008] model.layers.15.input_layernorm.weight -> blk.15.attn_norm.weight | F16 | [4096] model.layers.15.post_attention_layernorm.weight -> blk.15.ffn_norm.weight | F16 | [4096] model.layers.16.self_attn.q_proj.weight -> blk.16.attn_q.weight | F16 | [4096, 4096] model.layers.16.self_attn.k_proj.weight -> blk.16.attn_k.weight | F16 | [4096, 4096] model.layers.16.self_attn.v_proj.weight -> blk.16.attn_v.weight | F16 | [4096, 4096] model.layers.16.self_attn.o_proj.weight -> blk.16.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.16.attn_rot_embd model.layers.16.mlp.gate_proj.weight -> blk.16.ffn_gate.weight | F16 | [11008, 4096] model.layers.16.mlp.up_proj.weight -> blk.16.ffn_up.weight | F16 | [11008, 4096] model.layers.16.mlp.down_proj.weight -> blk.16.ffn_down.weight | F16 | [4096, 11008] model.layers.16.input_layernorm.weight -> blk.16.attn_norm.weight | F16 | [4096] model.layers.16.post_attention_layernorm.weight -> blk.16.ffn_norm.weight | F16 | [4096] model.layers.17.self_attn.q_proj.weight -> blk.17.attn_q.weight | F16 | [4096, 4096] model.layers.17.self_attn.k_proj.weight -> blk.17.attn_k.weight | F16 | [4096, 4096] model.layers.17.self_attn.v_proj.weight -> blk.17.attn_v.weight | F16 | [4096, 4096] model.layers.17.self_attn.o_proj.weight -> blk.17.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.17.attn_rot_embd model.layers.17.mlp.gate_proj.weight -> blk.17.ffn_gate.weight | F16 | [11008, 4096] model.layers.17.mlp.up_proj.weight -> blk.17.ffn_up.weight | F16 | [11008, 4096] model.layers.17.mlp.down_proj.weight -> blk.17.ffn_down.weight | F16 | [4096, 11008] model.layers.17.input_layernorm.weight -> blk.17.attn_norm.weight | F16 | [4096] model.layers.17.post_attention_layernorm.weight -> blk.17.ffn_norm.weight | F16 | [4096] model.layers.18.self_attn.q_proj.weight -> blk.18.attn_q.weight | F16 | [4096, 4096] model.layers.18.self_attn.k_proj.weight -> blk.18.attn_k.weight | F16 | [4096, 4096] model.layers.18.self_attn.v_proj.weight -> blk.18.attn_v.weight | F16 | [4096, 4096] model.layers.18.self_attn.o_proj.weight -> blk.18.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.18.attn_rot_embd model.layers.18.mlp.gate_proj.weight -> blk.18.ffn_gate.weight | F16 | [11008, 4096] model.layers.18.mlp.up_proj.weight -> blk.18.ffn_up.weight | F16 | [11008, 4096] model.layers.18.mlp.down_proj.weight -> blk.18.ffn_down.weight | F16 | [4096, 11008] model.layers.18.input_layernorm.weight -> blk.18.attn_norm.weight | F16 | [4096] model.layers.18.post_attention_layernorm.weight -> blk.18.ffn_norm.weight | F16 | [4096] model.layers.19.self_attn.q_proj.weight -> blk.19.attn_q.weight | F16 | [4096, 4096] model.layers.19.self_attn.k_proj.weight -> blk.19.attn_k.weight | F16 | [4096, 4096] model.layers.19.self_attn.v_proj.weight -> blk.19.attn_v.weight | F16 | [4096, 4096] model.layers.19.self_attn.o_proj.weight -> blk.19.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.19.attn_rot_embd model.layers.19.mlp.gate_proj.weight -> blk.19.ffn_gate.weight | F16 | [11008, 4096] model.layers.19.mlp.up_proj.weight -> blk.19.ffn_up.weight | F16 | [11008, 4096] model.layers.19.mlp.down_proj.weight -> blk.19.ffn_down.weight | F16 | [4096, 11008] model.layers.19.input_layernorm.weight -> blk.19.attn_norm.weight | F16 | [4096] model.layers.19.post_attention_layernorm.weight -> blk.19.ffn_norm.weight | F16 | [4096] model.layers.20.self_attn.q_proj.weight -> blk.20.attn_q.weight | F16 | [4096, 4096] model.layers.20.self_attn.k_proj.weight -> blk.20.attn_k.weight | F16 | [4096, 4096] model.layers.20.self_attn.v_proj.weight -> blk.20.attn_v.weight | F16 | [4096, 4096] model.layers.20.self_attn.o_proj.weight -> blk.20.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.20.attn_rot_embd model.layers.20.mlp.gate_proj.weight -> blk.20.ffn_gate.weight | F16 | [11008, 4096] model.layers.20.mlp.up_proj.weight -> blk.20.ffn_up.weight | F16 | [11008, 4096] model.layers.20.mlp.down_proj.weight -> blk.20.ffn_down.weight | F16 | [4096, 11008] model.layers.20.input_layernorm.weight -> blk.20.attn_norm.weight | F16 | [4096] model.layers.20.post_attention_layernorm.weight -> blk.20.ffn_norm.weight | F16 | [4096] model.layers.21.self_attn.q_proj.weight -> blk.21.attn_q.weight | F16 | [4096, 4096] model.layers.21.self_attn.k_proj.weight -> blk.21.attn_k.weight | F16 | [4096, 4096] model.layers.21.self_attn.v_proj.weight -> blk.21.attn_v.weight | F16 | [4096, 4096] model.layers.21.self_attn.o_proj.weight -> blk.21.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.21.attn_rot_embd model.layers.21.mlp.gate_proj.weight -> blk.21.ffn_gate.weight | F16 | [11008, 4096] model.layers.21.mlp.up_proj.weight -> blk.21.ffn_up.weight | F16 | [11008, 4096] model.layers.21.mlp.down_proj.weight -> blk.21.ffn_down.weight | F16 | [4096, 11008] model.layers.21.input_layernorm.weight -> blk.21.attn_norm.weight | F16 | [4096] model.layers.21.post_attention_layernorm.weight -> blk.21.ffn_norm.weight | F16 | [4096] model.layers.22.self_attn.q_proj.weight -> blk.22.attn_q.weight | F16 | [4096, 4096] model.layers.22.self_attn.k_proj.weight -> blk.22.attn_k.weight | F16 | [4096, 4096] model.layers.22.self_attn.v_proj.weight -> blk.22.attn_v.weight | F16 | [4096, 4096] model.layers.22.self_attn.o_proj.weight -> blk.22.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.22.attn_rot_embd model.layers.22.mlp.gate_proj.weight -> blk.22.ffn_gate.weight | F16 | [11008, 4096] model.layers.22.mlp.up_proj.weight -> blk.22.ffn_up.weight | F16 | [11008, 4096] model.layers.22.mlp.down_proj.weight -> blk.22.ffn_down.weight | F16 | [4096, 11008] model.layers.22.input_layernorm.weight -> blk.22.attn_norm.weight | F16 | [4096] model.layers.22.post_attention_layernorm.weight -> blk.22.ffn_norm.weight | F16 | [4096] model.layers.23.self_attn.q_proj.weight -> blk.23.attn_q.weight | F16 | [4096, 4096] model.layers.23.self_attn.k_proj.weight -> blk.23.attn_k.weight | F16 | [4096, 4096] model.layers.23.self_attn.v_proj.weight -> blk.23.attn_v.weight | F16 | [4096, 4096] model.layers.23.self_attn.o_proj.weight -> blk.23.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.23.attn_rot_embd model.layers.23.mlp.gate_proj.weight -> blk.23.ffn_gate.weight | F16 | [11008, 4096] model.layers.23.mlp.up_proj.weight -> blk.23.ffn_up.weight | F16 | [11008, 4096] model.layers.23.mlp.down_proj.weight -> blk.23.ffn_down.weight | F16 | [4096, 11008] model.layers.23.input_layernorm.weight -> blk.23.attn_norm.weight | F16 | [4096] model.layers.23.post_attention_layernorm.weight -> blk.23.ffn_norm.weight | F16 | [4096] model.layers.24.self_attn.q_proj.weight -> blk.24.attn_q.weight | F16 | [4096, 4096] model.layers.24.self_attn.k_proj.weight -> blk.24.attn_k.weight | F16 | [4096, 4096] model.layers.24.self_attn.v_proj.weight -> blk.24.attn_v.weight | F16 | [4096, 4096] model.layers.24.self_attn.o_proj.weight -> blk.24.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.24.attn_rot_embd model.layers.24.mlp.gate_proj.weight -> blk.24.ffn_gate.weight | F16 | [11008, 4096] model.layers.24.mlp.up_proj.weight -> blk.24.ffn_up.weight | F16 | [11008, 4096] model.layers.24.mlp.down_proj.weight -> blk.24.ffn_down.weight | F16 | [4096, 11008] model.layers.24.input_layernorm.weight -> blk.24.attn_norm.weight | F16 | [4096] model.layers.24.post_attention_layernorm.weight -> blk.24.ffn_norm.weight | F16 | [4096] model.layers.25.self_attn.q_proj.weight -> blk.25.attn_q.weight | F16 | [4096, 4096] model.layers.25.self_attn.k_proj.weight -> blk.25.attn_k.weight | F16 | [4096, 4096] model.layers.25.self_attn.v_proj.weight -> blk.25.attn_v.weight | F16 | [4096, 4096] model.layers.25.self_attn.o_proj.weight -> blk.25.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.25.attn_rot_embd model.layers.25.mlp.gate_proj.weight -> blk.25.ffn_gate.weight | F16 | [11008, 4096] model.layers.25.mlp.up_proj.weight -> blk.25.ffn_up.weight | F16 | [11008, 4096] model.layers.25.mlp.down_proj.weight -> blk.25.ffn_down.weight | F16 | [4096, 11008] model.layers.25.input_layernorm.weight -> blk.25.attn_norm.weight | F16 | [4096] model.layers.25.post_attention_layernorm.weight -> blk.25.ffn_norm.weight | F16 | [4096] model.layers.26.self_attn.q_proj.weight -> blk.26.attn_q.weight | F16 | [4096, 4096] model.layers.26.self_attn.k_proj.weight -> blk.26.attn_k.weight | F16 | [4096, 4096] model.layers.26.self_attn.v_proj.weight -> blk.26.attn_v.weight | F16 | [4096, 4096] model.layers.26.self_attn.o_proj.weight -> blk.26.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.26.attn_rot_embd model.layers.26.mlp.gate_proj.weight -> blk.26.ffn_gate.weight | F16 | [11008, 4096] model.layers.26.mlp.up_proj.weight -> blk.26.ffn_up.weight | F16 | [11008, 4096] model.layers.26.mlp.down_proj.weight -> blk.26.ffn_down.weight | F16 | [4096, 11008] model.layers.26.input_layernorm.weight -> blk.26.attn_norm.weight | F16 | [4096] model.layers.26.post_attention_layernorm.weight -> blk.26.ffn_norm.weight | F16 | [4096] model.layers.27.self_attn.q_proj.weight -> blk.27.attn_q.weight | F16 | [4096, 4096] model.layers.27.self_attn.k_proj.weight -> blk.27.attn_k.weight | F16 | [4096, 4096] model.layers.27.self_attn.v_proj.weight -> blk.27.attn_v.weight | F16 | [4096, 4096] model.layers.27.self_attn.o_proj.weight -> blk.27.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.27.attn_rot_embd model.layers.27.mlp.gate_proj.weight -> blk.27.ffn_gate.weight | F16 | [11008, 4096] model.layers.27.mlp.up_proj.weight -> blk.27.ffn_up.weight | F16 | [11008, 4096] model.layers.27.mlp.down_proj.weight -> blk.27.ffn_down.weight | F16 | [4096, 11008] model.layers.27.input_layernorm.weight -> blk.27.attn_norm.weight | F16 | [4096] model.layers.27.post_attention_layernorm.weight -> blk.27.ffn_norm.weight | F16 | [4096] model.layers.28.self_attn.q_proj.weight -> blk.28.attn_q.weight | F16 | [4096, 4096] model.layers.28.self_attn.k_proj.weight -> blk.28.attn_k.weight | F16 | [4096, 4096] model.layers.28.self_attn.v_proj.weight -> blk.28.attn_v.weight | F16 | [4096, 4096] model.layers.28.self_attn.o_proj.weight -> blk.28.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.28.attn_rot_embd model.layers.28.mlp.gate_proj.weight -> blk.28.ffn_gate.weight | F16 | [11008, 4096] model.layers.28.mlp.up_proj.weight -> blk.28.ffn_up.weight | F16 | [11008, 4096] model.layers.28.mlp.down_proj.weight -> blk.28.ffn_down.weight | F16 | [4096, 11008] model.layers.28.input_layernorm.weight -> blk.28.attn_norm.weight | F16 | [4096] model.layers.28.post_attention_layernorm.weight -> blk.28.ffn_norm.weight | F16 | [4096] model.layers.29.self_attn.q_proj.weight -> blk.29.attn_q.weight | F16 | [4096, 4096] model.layers.29.self_attn.k_proj.weight -> blk.29.attn_k.weight | F16 | [4096, 4096] model.layers.29.self_attn.v_proj.weight -> blk.29.attn_v.weight | F16 | [4096, 4096] model.layers.29.self_attn.o_proj.weight -> blk.29.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.29.attn_rot_embd model.layers.29.mlp.gate_proj.weight -> blk.29.ffn_gate.weight | F16 | [11008, 4096] model.layers.29.mlp.up_proj.weight -> blk.29.ffn_up.weight | F16 | [11008, 4096] model.layers.29.mlp.down_proj.weight -> blk.29.ffn_down.weight | F16 | [4096, 11008] model.layers.29.input_layernorm.weight -> blk.29.attn_norm.weight | F16 | [4096] model.layers.29.post_attention_layernorm.weight -> blk.29.ffn_norm.weight | F16 | [4096] model.layers.30.self_attn.q_proj.weight -> blk.30.attn_q.weight | F16 | [4096, 4096] model.layers.30.self_attn.k_proj.weight -> blk.30.attn_k.weight | F16 | [4096, 4096] model.layers.30.self_attn.v_proj.weight -> blk.30.attn_v.weight | F16 | [4096, 4096] model.layers.30.self_attn.o_proj.weight -> blk.30.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.30.attn_rot_embd model.layers.30.mlp.gate_proj.weight -> blk.30.ffn_gate.weight | F16 | [11008, 4096] model.layers.30.mlp.up_proj.weight -> blk.30.ffn_up.weight | F16 | [11008, 4096] model.layers.30.mlp.down_proj.weight -> blk.30.ffn_down.weight | F16 | [4096, 11008] model.layers.30.input_layernorm.weight -> blk.30.attn_norm.weight | F16 | [4096] model.layers.30.post_attention_layernorm.weight -> blk.30.ffn_norm.weight | F16 | [4096] model.layers.31.self_attn.q_proj.weight -> blk.31.attn_q.weight | F16 | [4096, 4096] model.layers.31.self_attn.k_proj.weight -> blk.31.attn_k.weight | F16 | [4096, 4096] model.layers.31.self_attn.v_proj.weight -> blk.31.attn_v.weight | F16 | [4096, 4096] model.layers.31.self_attn.o_proj.weight -> blk.31.attn_output.weight | F16 | [4096, 4096] skipping tensor blk.31.attn_rot_embd model.layers.31.mlp.gate_proj.weight -> blk.31.ffn_gate.weight | F16 | [11008, 4096] model.layers.31.mlp.up_proj.weight -> blk.31.ffn_up.weight | F16 | [11008, 4096] model.layers.31.mlp.down_proj.weight -> blk.31.ffn_down.weight | F16 | [4096, 11008] model.layers.31.input_layernorm.weight -> blk.31.attn_norm.weight | F16 | [4096] model.layers.31.post_attention_layernorm.weight -> blk.31.ffn_norm.weight | F16 | [4096] model.norm.weight -> output_norm.weight | F16 | [4096] lm_head.weight -> output.weight | F16 | [32000, 4096] Writing llama-2-7b-aiopsfinetunedv2-q8_0-gguf, format 7 Ignoring added_tokens.json since model matches vocab size without it. gguf: This GGUF file is for Little Endian only gguf: Setting special token type bos to 1 gguf: Setting special token type eos to 2 gguf: Setting special token type unk to 0 gguf: Setting special token type pad to 0 Traceback (most recent call last): File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 1483, in main() File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 1477, in main OutputFile.write_all(outfile, ftype, params, model, vocab, special_vocab, File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 1144, in write_all for i, ((name, lazy_tensor), ndarray) in enumerate(zip(model.items(), ndarrays)): File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 918, in bounded_parallel_map futures.append(executor.submit(func, next(iterable))) ^^^^^^^^^^^^^^ File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 924, in bounded_parallel_map result = futures.pop(0).result() ^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/lib64/python3.12/concurrent/futures/_base.py", line 449, in result return self.__get_result() ^^^^^^^^^^^^^^^^^^^ File "/usr/lib64/python3.12/concurrent/futures/_base.py", line 401, in __get_result raise self._exception File "/usr/lib64/python3.12/concurrent/futures/thread.py", line 58, in run result = self.fn(*self.args, **self.kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 1101, in do_item tensor = lazy_tensor.load().to_ggml() ^^^^^^^^^^^^^^^^^^ File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 649, in load ret = self._load() ^^^^^^^^^^^^ File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 659, in load return self.load().astype(data_type) ^^^^^^^^^^^ File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 649, in load ret = self._load() ^^^^^^^^^^^^ File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 737, in load return lazy_tensor.load().permute(n_head, n_head_kv) ^^^^^^^^^^^^^^^^^^ File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 649, in load ret = self._load() ^^^^^^^^^^^^ File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 809, in load return UnquantizedTensor(storage.load(storage_offset, elm_count).reshape(size)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 796, in load data = fp.read(size) ^^^^^^^^^^^^^ File "/usr/lib64/python3.12/zipfile/init.py", line 979, in read data = self._read1(n) ^^^^^^^^^^^^^^ File "/usr/lib64/python3.12/zipfile/init.py", line 1069, in _read1 self._update_crc(data) File "/usr/lib64/python3.12/zipfile/init.py", line 997, in _update_crc raise BadZipFile("Bad CRC-32 for file %r" % self.name) zipfile.BadZipFile: Bad CRC-32 for file 'pytorch_model-00001-of-00002/data/2'

It did start with the requirments.txt not installing. I had to replace the ~= with >= to make it install. I am using a clean virtual python environement.

aspanner avatar Feb 17 '24 01:02 aspanner

It did start with the requirments.txt not installing. I had to replace the ~= with >= to make it install. I am using a clean virtual python environement.

This is due to ImpImporter being removed recently from pip (the Python thread where ImpImporter is deprecated). The zipfile issue also seems to be 3.12 specific. Switching to 3.10 in my environment fixed the issue for me.

JaCraig avatar Feb 22 '24 15:02 JaCraig

Thanks @JaCraig - Can confirm that switching from 3.12 to 3.10 fixed the issue for me as well.

wdavo avatar Mar 04 '24 10:03 wdavo

Facing the same issue. I don't want to downgrade my python just for this. I think llama.cpp should support Python 3.12. The latest pytorch version (2.2.1) now also supports Python 3.12 (https://github.com/pytorch/pytorch/issues/110436#issuecomment-1957259894) so it shouldn't be a problem to support it.

harshil21 avatar Mar 09 '24 00:03 harshil21

Same issue here, on windows 10 PC: Traceback (most recent call last): File "D:\eosusa\cpp\llama.cpp\convert.py", line 1486, in main() File "D:\eosusa\cpp\llama.cpp\convert.py", line 1480, in main OutputFile.write_all(outfile, ftype, params, model, vocab, special_vocab, File "D:\eosusa\cpp\llama.cpp\convert.py", line 1162, in write_all of.write_tensor_data(ftype, model, concurrency) File "D:\eosusa\cpp\llama.cpp\convert.py", line 1093, in write_tensor_data for i, ((name, lazy_tensor), ndarray) in enumerate(zip(model.items(), ndarrays)): File "D:\eosusa\cpp\llama.cpp\convert.py", line 941, in bounded_parallel_map result = futures.pop(0).result() ^^^^^^^^^^^^^^^^^^^^^^^ File "C:\ProgramData\miniconda3\Lib\concurrent\futures_base.py", line 449, in result return self.__get_result() ^^^^^^^^^^^^^^^^^^^ File "C:\ProgramData\miniconda3\Lib\concurrent\futures_base.py", line 401, in __get_result raise self._exception File "C:\ProgramData\miniconda3\Lib\concurrent\futures\thread.py", line 58, in run result = self.fn(*self.args, **self.kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\eosusa\cpp\llama.cpp\convert.py", line 1126, in do_item tensor = lazy_tensor.load().to_ggml() ^^^^^^^^^^^^^^^^^^ File "D:\eosusa\cpp\llama.cpp\convert.py", line 666, in load ret = self._load() ^^^^^^^^^^^^ File "D:\eosusa\cpp\llama.cpp\convert.py", line 676, in load return self.load().astype(data_type) ^^^^^^^^^^^ File "D:\eosusa\cpp\llama.cpp\convert.py", line 666, in load ret = self.load() ^^^^^^^^^^^^ File "D:\eosusa\cpp\llama.cpp\convert.py", line 826, in load return UnquantizedTensor(storage.load(storage_offset, elm_count).reshape(size)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\eosusa\cpp\llama.cpp\convert.py", line 810, in load fp = self.zip_file.open(info) ^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\ProgramData\miniconda3\Lib\zipfile_init.py", line 1643, in open raise BadZipFile(f"Overlapped entries: {zinfo.orig_filename!r} (possible zip bomb)") zipfile.BadZipFile: Overlapped entries: 'consolidated.00/data/73' (possible zip bomb)

python --version Python 3.12.2

All requriements.txt installed.

MarcoLv412 avatar Mar 25 '24 22:03 MarcoLv412

the same issue on macos: data = self._read1(n) File "/usr/local/Cellar/[email protected]/3.9.13_2/Frameworks/Python.framework/Versions/3.9/lib/python3.9/zipfile.py", line 1014, in _read1 self._update_crc(data) File "/usr/local/Cellar/[email protected]/3.9.13_2/Frameworks/Python.framework/Versions/3.9/lib/python3.9/zipfile.py", line 942, in _update_crc raise BadZipFile("Bad CRC-32 for file %r" % self.name) zipfile.BadZipFile: Bad CRC-32 for file 'pytorch_model-00001-of-00002/data/144'

llmama.cpp version: commit 4e96a812b3ce7322a29a3008db2ed73d9087b176 (HEAD -> master, origin/master, origin/HEAD) Author: Anas Ahouzi [email protected] Date: Tue Apr 23 02:53:18 2024 +0200

python3.10

ifoxhz avatar Apr 28 '24 01:04 ifoxhz

It did start with the requirments.txt not installing. I had to replace the ~= with >= to make it install. I am using a clean virtual python environement.

This is due to ImpImporter being removed recently from pip (the Python thread where ImpImporter is deprecated). The zipfile issue also seems to be 3.12 specific. Switching to 3.10 in my environment fixed the issue for me.

Thanks @JaCraig Solved my problem.

Leonidas-Li avatar May 08 '24 09:05 Leonidas-Li

Same issue here, on windows 10 PC: Traceback (most recent call last): File "D:\eosusa\cpp\llama.cpp\convert.py", line 1486, in main() File "D:\eosusa\cpp\llama.cpp\convert.py", line 1480, in main OutputFile.write_all(outfile, ftype, params, model, vocab, special_vocab, File "D:\eosusa\cpp\llama.cpp\convert.py", line 1162, in write_all of.write_tensor_data(ftype, model, concurrency) File "D:\eosusa\cpp\llama.cpp\convert.py", line 1093, in write_tensor_data for i, ((name, lazy_tensor), ndarray) in enumerate(zip(model.items(), ndarrays)): File "D:\eosusa\cpp\llama.cpp\convert.py", line 941, in bounded_parallel_map result = futures.pop(0).result() ^^^^^^^^^^^^^^^^^^^^^^^ File "C:\ProgramData\miniconda3\Lib\concurrent\futures_base.py", line 449, in result return self.__get_result() ^^^^^^^^^^^^^^^^^^^ File "C:\ProgramData\miniconda3\Lib\concurrent\futures_base.py", line 401, in __get_result raise self._exception File "C:\ProgramData\miniconda3\Lib\concurrent\futures\thread.py", line 58, in run result = self.fn(*self.args, **self.kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\eosusa\cpp\llama.cpp\convert.py", line 1126, in do_item tensor = lazy_tensor.load().to_ggml() ^^^^^^^^^^^^^^^^^^ File "D:\eosusa\cpp\llama.cpp\convert.py", line 666, in load ret = self.load() ^^^^^^^^^^^^ File "D:\eosusa\cpp\llama.cpp\convert.py", line 676, in load return self.load().astype(data_type) ^^^^^^^^^^^ File "D:\eosusa\cpp\llama.cpp\convert.py", line 666, in load ret = self.load() ^^^^^^^^^^^^ File "D:\eosusa\cpp\llama.cpp\convert.py", line 826, in load return UnquantizedTensor(storage.load(storage_offset, elm_count).reshape(size)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\eosusa\cpp\llama.cpp\convert.py", line 810, in load fp = self.zip_file.open(info) ^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\ProgramData\miniconda3\Lib\zipfile__init.py", line 1643, in open raise BadZipFile(f"Overlapped entries: {zinfo.orig_filename!r} (possible zip bomb)") zipfile.BadZipFile: Overlapped entries: 'consolidated.00/data/73' (possible zip bomb)

python --version Python 3.12.2

All requriements.txt installed.

@MarcoLv412 Same error message here on Windows 11, python=3.8. Try adding "--concurrency=1" which works for me. Seems some problems in multi-thread zipfile reading.

QingtaoLi1 avatar May 20 '24 07:05 QingtaoLi1

@mofosyne This is a bug. It is python specific though. Add to docs for now?

Using safetensors is a valid workaround.

teleprint-me avatar May 20 '24 21:05 teleprint-me

This issue was closed because it has been inactive for 14 days since being marked as stale.

github-actions[bot] avatar Jul 06 '24 01:07 github-actions[bot]