orellavie1212
orellavie1212
Dear Support Team, I recently upgraded my pip libraries, including transformers, peft, accelerate, and bitsandbytes, to support 4-bit training as opposed to the original 8-bit training. After doing so, I...
### 🚀 The feature, motivation and pitch Bitsandbytes 4bit quantization support. I know many want that, and also it is discuused before and marked as unplaned, but after I looked...
Hey, After testing failure of fuzz.ratio at the detection of basic sim. match, I found that the newer ver of python-Levenshtein(0.22.0) makes the fuzz.ratio output wrong answer when installing it...
Hello, Wanted to quantize the model via awq after a merged qlora b&b nf4 mixtral moe. the error is: ``` self._search_best_scale(self.modules[i], **layer) File "/home/access/anaconda3/envs/sec_qlora_replicate/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 115, in decorate_context return func(*args,...
### Your current environment sagemaker ml.g5.12xlarge instance (4 instances of a10g 24gb) container is 763104351884.dkr.ecr.us-west-2.amazonaws.com/djl-inference:0.27.0-deepspeed0.12.6-cu121 from https://github.com/aws/deep-learning-containers/blob/master/available_images.md ### 🐛 Describe the bug from vllm import LLM, SamplingParams question = "what...