ABQ-LLM icon indicating copy to clipboard operation
ABQ-LLM copied to clipboard

No reduction in model size

Open Sekri0 opened this issue 4 months ago • 2 comments

I use this command to quantize llama2-7b-chat model, but the model size dosen't change. CUDA_VISIBLE_DEVICES=0 python3 main.py
--model /mnt/home/model/llama2-7b-chat-hf
--epochs 20 --output_dir ./log/llama2-7b-w2a8
--eval_ppl --wbits 2 --abits 8 --lwc --let
--tasks piqa,arc_easy,arc_challenge,boolq,hellaswag,winogrande
--real_quant
--save_dir /mnt/home/model/abq-llm/llama2-7b-w2a8

Sekri0 avatar Oct 01 '24 03:10 Sekri0