Marc Sun
Marc Sun
Hi @parasurama, thanks for reporting ! I'll have a look asap
Hi @parasurama, this happens because you changed `max_position_embeddings` attribute. This modified a lot of weights and the whole model needs to be retrained. For now, we don't support loading mismatched...
This happens because the default vocab_size of LlamaConfig is 32000 but llama v3 checkpoint have a vocab_size of 128256 but llama v2 checkpoint have a vocab_size of 32000. So by...
Hi @goelayu, this is expected since `with torch.device('meta')` also puts the buffers on the `meta` device. However, non persistant buffers are not saved in the `state_dict`. So, in the case...
Hi @jkterry1, thanks for this detailed report ! For 3. and 4. , let me know if you want to submit a PR to fix the logger message and split...
Closing this since the issue is solved !
I see that `transformers-all-latest-gpu` docker image is not being updated for the last two days since the [installation](https://github.com/huggingface/transformers/actions/runs/7924158967/job/21635225922) fails because of aqml library that requires python 3.10 at least and...
> I was able to run aqlm on python 3.8 no problem otherwise. I can replace the statement with an if-else statement and lower the requirement if necessary. Yes that...
Perfect ! I will wait for your PR to be merged then + release then if it doesn't take too much time. Please keep me updated ! Otherwise, I can...
I was able to build the image but I don't have the permission to push the image cc @ydshieh `#22 ERROR: failed to push huggingface/transformers-quantization-latest-gpu: push access denied, repository does...