Xingtong Liu
Xingtong Liu
My understanding is currently the repo provides 4bit only for inference but not finetuning. If this is the case, is there a plan for integrating QLoRA-style 4bit finetuning?
Whenever I try to load a QLoRA-merged Falcon 40B model, the error below happens. `error loading model: falcon.cpp: tensor 'lm_head.weight' is missing from model` The hacky way I did to...
https://github.com/facebookresearch/segment-anything/blob/main/segment_anything/modeling/prompt_encoder.py#L81-L85 If I read it correctly, the shape of points input is `[B, N, 2]`, where B is the batch size and N is the number of points per image....