johnysh
johnysh
The current model is unable to calculate the time spent on first token and rest tokens, can we add this msg ?
When ran InternLM2 inference , it reported errors as below: oneAPI :2024.0.1.46 ipex-llm: 2.1.0b2 transformers: 4.37.2 ,4.38.2 --------------------------------------------------------------------------- --------------------------------------------------------------------------- TypeError Traceback (most recent call last) Cell In[1], line 33 31...
I ran glm4 on MTL iGPU , and it reported this error:  oneAPI: l_BaseKit_p_2024.0.1.46_offline.sh My env as below: (notebook-zone) arc@arc:~/ipex-llm/python/llm/example/GPU/HuggingFace/LLM/glm4$ pip list Package Version ----------------------------- ------------------ accelerate 0.23.0 addict...