alpaca-lora
alpaca-lora copied to clipboard
CUDA out of memory : I am using Colab T4 GPU
!python generate.py
You are using the default legacy behaviour of the <class 'transformers.models.llama.tokenization_llama.LlamaTokenizer'>. This is expected, and simply means that the legacy
(previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set legacy=False
. This should only be set if you understand what it means, and thouroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565
Loading checkpoint shards: 100% 33/33 [01:13<00:00, 2.22s/it]
Instruction: Tell me about alpacas.
/usr/local/lib/python3.10/dist-packages/transformers/generation/configuration_utils.py:381: UserWarning: do_sample
is set to False
. However, temperature
is set to 0.1
-- this flag is only used in sample-based generation modes. You should set do_sample=True
or unset temperature
. This was detected when initializing the generation config instance, which means the corresponding file may hold incorrect parameterization and should be fixed.
warnings.warn(
/usr/local/lib/python3.10/dist-packages/transformers/generation/configuration_utils.py:386: UserWarning: do_sample
is set to False
. However, top_p
is set to 0.75
-- this flag is only used in sample-based generation modes. You should set do_sample=True
or unset top_p
. This was detected when initializing the generation config instance, which means the corresponding file may hold incorrect parameterization and should be fixed.
warnings.warn(
/usr/local/lib/python3.10/dist-packages/transformers/generation/configuration_utils.py:396: UserWarning: do_sample
is set to False
. However, top_k
is set to 40
-- this flag is only used in sample-based generation modes. You should set do_sample=True
or unset top_k
. This was detected when initializing the generation config instance, which means the corresponding file may hold incorrect parameterization and should be fixed.
warnings.warn(
/usr/local/lib/python3.10/dist-packages/transformers/generation/configuration_utils.py:381: UserWarning: do_sample
is set to False
. However, temperature
is set to 0.1
-- this flag is only used in sample-based generation modes. You should set do_sample=True
or unset temperature
.
warnings.warn(
/usr/local/lib/python3.10/dist-packages/transformers/generation/configuration_utils.py:386: UserWarning: do_sample
is set to False
. However, top_p
is set to 0.75
-- this flag is only used in sample-based generation modes. You should set do_sample=True
or unset top_p
.
warnings.warn(
/usr/local/lib/python3.10/dist-packages/transformers/generation/configuration_utils.py:396: UserWarning: do_sample
is set to False
. However, top_k
is set to 40
-- this flag is only used in sample-based generation modes. You should set do_sample=True
or unset top_k
.
warnings.warn(
Traceback (most recent call last):
File "/content/generate.py", line 150, in
I came across this error too,do you know how to solve it ^o^
No, I have not yet solved this issue.