ABQ-LLM
ABQ-LLM copied to clipboard
No reduction in model size
I use this command to quantize llama2-7b-chat model, but the model size dosen't change.
CUDA_VISIBLE_DEVICES=0 python3 main.py
--model /mnt/home/model/llama2-7b-chat-hf
--epochs 20 --output_dir ./log/llama2-7b-w2a8
--eval_ppl --wbits 2 --abits 8 --lwc --let
--tasks piqa,arc_easy,arc_challenge,boolq,hellaswag,winogrande
--real_quant
--save_dir /mnt/home/model/abq-llm/llama2-7b-w2a8