exo
exo copied to clipboard
Getting Killed
The last meaningful output:
ram used: 5.67 GB, tok_embeddings.weight : 99%|▉| 290/292 [00:25<00:00, 11.3
ram used: 6.72 GB, output.weight : 100%|▉| 291/292 [00:25<00:00, 11.4
ram used: 6.72 GB, freqs_cis : 100%|█| 292/292 [00:25<00:00, 11.4
ram used: 6.72 GB, freqs_cis : 100%|█| 292/292 [00:25<00:00, 11.4
loaded weights in 25547.80 ms, 6.72 GB loaded at 0.26 GB/s
Checking if local path exists to load tokenizer from local local_path=None
Trying AutoProcessor for
/home/ffamax/.cache/huggingface/hub/models--mlabonne--Meta-Llama-3.1-8B-Instruct-abliterated/snapshots/368
c8ed94ce4c986e7b9ca5c159651ef753908ce
get_allow_patterns weight_map={'lm_head.weight': 'model-00004-of-00004.safetensors',
'model.embed_tokens.weight': 'model-00001-of-00004.safetensors', 'model.layers.0.input_layernorm.weight':
'model-00001-of-00004.safetensors', 'model.layers.0.mlp.down_proj.weight':
'model-00001-of-00004.safetensors', 'model.layers.0.mlp.gate_proj.weight':
... skipped ...
'model-00002-of-00004.safetensors', 'model.layers.9.self_attn.o_proj.weight':
'model-00002-of-00004.safetensors', 'model.layers.9.self_attn.q_proj.weight':
'model-00002-of-00004.safetensors', 'model.layers.9.self_attn.v_proj.weight':
'model-00002-of-00004.safetensors', 'model.norm.weight': 'model-00004-of-00004.safetensors'}
shard=Shard(model_id='mlabonne/Meta-Llama-3.1-8B-Instruct-abliterated', start_layer=0, end_layer=12,
n_layers=32) shard_specific_patterns={'model-00002-of-00004.safetensors',
'model-00001-of-00004.safetensors'}
Received request: GET /v1/download/progress
update_peers: added=[] removed=[] updated=[]
unchanged=[<exo.networking.grpc.grpc_peer_handle.GRPCPeerHandle object at 0x7f73ae7afd00>,
<exo.networking.grpc.grpc_peer_handle.GRPCPeerHandle object at 0x7f73ae7afb20>] to_disconnect=[]
to_connect=[]
did_peers_change=False
Received request: GET /v1/download/progress
update_peers: added=[] removed=[] updated=[]
unchanged=[<exo.networking.grpc.grpc_peer_handle.GRPCPeerHandle object at 0x7f73ae7afd00>,
<exo.networking.grpc.grpc_peer_handle.GRPCPeerHandle object at 0x7f73ae7afb20>] to_disconnect=[]
to_connect=[]
did_peers_change=False
Received request: GET /v1/download/progress
update_peers: added=[] removed=[] updated=[]
unchanged=[<exo.networking.grpc.grpc_peer_handle.GRPCPeerHandle object at 0x7f73ae7afd00>,
<exo.networking.grpc.grpc_peer_handle.GRPCPeerHandle object at 0x7f73ae7afb20>] to_disconnect=[]
to_connect=[]
did_peers_change=False
Received request: GET /v1/download/progress
Excluded model param keys for shard=Shard(model_id='mlabonne/Meta-Llama-3.1-8B-Instruct-abliterated',
start_layer=0, end_layer=12, n_layers=32): ['lm_head.weight', 'model.layers.13.input_layernorm.weight',
'model.layers.13.mlp.down_proj.weight', 'model.layers.13.mlp.gate_proj.weight',
'model.layers.13.mlp.up_proj.weight', 'model.layers.13.post_attention_layernorm.weight',
... skipped ...
'model.layers.30.self_attn.v_proj.weight', 'model.layers.31.input_layernorm.weight',
'model.layers.31.mlp.down_proj.weight', 'model.layers.31.mlp.gate_proj.weight',
'model.layers.31.mlp.up_proj.weight', 'model.layers.31.post_attention_layernorm.weight',
'model.layers.31.self_attn.k_proj.weight', 'model.layers.31.self_attn.o_proj.weight',
'model.layers.31.self_attn.q_proj.weight', 'model.layers.31.self_attn.v_proj.weight', 'model.norm.weight']
0%| | 0/292 [00:00<?, ?it/s]
update_peers: added=[] removed=[] updated=[]
unchanged=[<exo.networking.grpc.grpc_peer_handle.GRPCPeerHandle object at 0x7f73ae7afd00>,
<exo.networking.grpc.grpc_peer_handle.GRPCPeerHandle object at 0x7f73ae7afb20>] to_disconnect=[]
to_connect=[]
did_peers_change=False
ram used: 6.72 GB, layers.0.attention.wq.weight : 0%| | 1/292 [00:00<02:22, 2.04i
ram used: 6.76 GB, layers.0.attention.wk.weight : 1%| | 2/292 [00:00<01:17, 3.73i
ram used: 6.76 GB, layers.0.attention.wv.weight : 1%| | 3/292 [00:00<00:56, 5.09i
update_peers: added=[] removed=[] updated=[]
unchanged=[<exo.networking.grpc.grpc_peer_handle.GRPCPeerHandle object at 0x7f73ae7afd00>,
<exo.networking.grpc.grpc_peer_handle.GRPCPeerHandle object at 0x7f73ae7afb20>] to_disconnect=[]
to_connect=[]
did_peers_change=False
ram used: 6.77 GB, layers.0.attention.wo.weight : 1%| | 4/292 [00:06<08:00, 0.60i
Received request: GET /v1/download/progress
Killed
Any ideas?