werruww
werruww
options.lang = ['eng', 'ara'] 2 options.lang = ['en', 'ar']
https://github.com/werruww/succ-docling-/blob/main/suc_docling%20(2).ipynb
https://github.com/lyogavin/airllm/issues/192
max_new_tokens=xxxxxxxxxxxxxxxxxxxxxx, The higher the number, the more load on the graphics unit and the longer the time. generation_output = model.generate( input_tokens['input_ids'].cuda(), max_new_tokens=12, use_cache=True, return_dict_in_generate=True) model.tokenizer.decode(generation_output.sequences[0])
it run on colab t4
from airllm import AutoModel MAX_LENGTH = 128 # could use hugging face model repo id: #model = AutoModel.from_pretrained("Qwen/Qwen-7B", profiling_mode=True) from airllm import AutoModel MAX_LENGTH = 128 # could use hugging...
How do I prevent the question "Who invented the electric light bulb? A. Thomas Edison" from being repeated and the answer being "A. Thomas Edison" only?
How do I prevent the question from being repeated?
echo=True Is there an instance of echo=True in airllm
Echo the prompt back in the output