exo icon indicating copy to clipboard operation
exo copied to clipboard

Getting Killed

Open FFAMax opened this issue 3 months ago • 2 comments

The last meaningful output:

ram used:  5.67 GB, tok_embeddings.weight                             :  99%|▉| 290/292 [00:25<00:00, 11.3
ram used:  6.72 GB, output.weight                                     : 100%|▉| 291/292 [00:25<00:00, 11.4
ram used:  6.72 GB, freqs_cis                                         : 100%|█| 292/292 [00:25<00:00, 11.4
ram used:  6.72 GB, freqs_cis                                         : 100%|█| 292/292 [00:25<00:00, 11.4
loaded weights in 25547.80 ms, 6.72 GB loaded at 0.26 GB/s
Checking if local path exists to load tokenizer from local local_path=None
Trying AutoProcessor for
/home/ffamax/.cache/huggingface/hub/models--mlabonne--Meta-Llama-3.1-8B-Instruct-abliterated/snapshots/368
c8ed94ce4c986e7b9ca5c159651ef753908ce
get_allow_patterns weight_map={'lm_head.weight': 'model-00004-of-00004.safetensors',
'model.embed_tokens.weight': 'model-00001-of-00004.safetensors', 'model.layers.0.input_layernorm.weight':
'model-00001-of-00004.safetensors', 'model.layers.0.mlp.down_proj.weight':
'model-00001-of-00004.safetensors', 'model.layers.0.mlp.gate_proj.weight':

... skipped ...

'model-00002-of-00004.safetensors', 'model.layers.9.self_attn.o_proj.weight':
'model-00002-of-00004.safetensors', 'model.layers.9.self_attn.q_proj.weight':
'model-00002-of-00004.safetensors', 'model.layers.9.self_attn.v_proj.weight':
'model-00002-of-00004.safetensors', 'model.norm.weight': 'model-00004-of-00004.safetensors'}
shard=Shard(model_id='mlabonne/Meta-Llama-3.1-8B-Instruct-abliterated', start_layer=0, end_layer=12,
n_layers=32) shard_specific_patterns={'model-00002-of-00004.safetensors',
'model-00001-of-00004.safetensors'}
Received request: GET /v1/download/progress
update_peers: added=[] removed=[] updated=[]
unchanged=[<exo.networking.grpc.grpc_peer_handle.GRPCPeerHandle object at 0x7f73ae7afd00>,
<exo.networking.grpc.grpc_peer_handle.GRPCPeerHandle object at 0x7f73ae7afb20>] to_disconnect=[]
to_connect=[]
did_peers_change=False
Received request: GET /v1/download/progress
update_peers: added=[] removed=[] updated=[]
unchanged=[<exo.networking.grpc.grpc_peer_handle.GRPCPeerHandle object at 0x7f73ae7afd00>,
<exo.networking.grpc.grpc_peer_handle.GRPCPeerHandle object at 0x7f73ae7afb20>] to_disconnect=[]
to_connect=[]
did_peers_change=False
Received request: GET /v1/download/progress
update_peers: added=[] removed=[] updated=[]
unchanged=[<exo.networking.grpc.grpc_peer_handle.GRPCPeerHandle object at 0x7f73ae7afd00>,
<exo.networking.grpc.grpc_peer_handle.GRPCPeerHandle object at 0x7f73ae7afb20>] to_disconnect=[]
to_connect=[]
did_peers_change=False
Received request: GET /v1/download/progress
Excluded model param keys for shard=Shard(model_id='mlabonne/Meta-Llama-3.1-8B-Instruct-abliterated',
start_layer=0, end_layer=12, n_layers=32): ['lm_head.weight', 'model.layers.13.input_layernorm.weight',
'model.layers.13.mlp.down_proj.weight', 'model.layers.13.mlp.gate_proj.weight',
'model.layers.13.mlp.up_proj.weight', 'model.layers.13.post_attention_layernorm.weight',

... skipped ...

'model.layers.30.self_attn.v_proj.weight', 'model.layers.31.input_layernorm.weight',
'model.layers.31.mlp.down_proj.weight', 'model.layers.31.mlp.gate_proj.weight',
'model.layers.31.mlp.up_proj.weight', 'model.layers.31.post_attention_layernorm.weight',
'model.layers.31.self_attn.k_proj.weight', 'model.layers.31.self_attn.o_proj.weight',
'model.layers.31.self_attn.q_proj.weight', 'model.layers.31.self_attn.v_proj.weight', 'model.norm.weight']
  0%|                                                                             | 0/292 [00:00<?, ?it/s]
update_peers: added=[] removed=[] updated=[]
unchanged=[<exo.networking.grpc.grpc_peer_handle.GRPCPeerHandle object at 0x7f73ae7afd00>,
<exo.networking.grpc.grpc_peer_handle.GRPCPeerHandle object at 0x7f73ae7afb20>] to_disconnect=[]
to_connect=[]
did_peers_change=False
ram used:  6.72 GB, layers.0.attention.wq.weight                      :   0%| | 1/292 [00:00<02:22,  2.04i
ram used:  6.76 GB, layers.0.attention.wk.weight                      :   1%| | 2/292 [00:00<01:17,  3.73i
ram used:  6.76 GB, layers.0.attention.wv.weight                      :   1%| | 3/292 [00:00<00:56,  5.09i
update_peers: added=[] removed=[] updated=[]
unchanged=[<exo.networking.grpc.grpc_peer_handle.GRPCPeerHandle object at 0x7f73ae7afd00>,
<exo.networking.grpc.grpc_peer_handle.GRPCPeerHandle object at 0x7f73ae7afb20>] to_disconnect=[]
to_connect=[]
did_peers_change=False
ram used:  6.77 GB, layers.0.attention.wo.weight                      :   1%| | 4/292 [00:06<08:00,  0.60i
Received request: GET /v1/download/progress
Killed

Any ideas?

FFAMax avatar Nov 03 '24 06:11 FFAMax