Results 9 issues of JamieVC

I followed installation guide https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Overview/install_gpu.html in order to run IPEX-LLM on iGPU of Meteor Lake with Windows OS platform. **---Steps to set environment ---** (llm) PS C:\Users\S54 PR> pip install...

user issue

Hi, I download langchain-chatchat and follow README.md. After that, putting two files to ingest. ( one is PDF, and the other is txt) ![image](https://github.com/intel-analytics/ipex-llm/assets/59042153/d579ea16-7739-41e4-9fc9-c8768336e711) It shows errors. ![image](https://github.com/intel-analytics/ipex-llm/assets/59042153/1fb693ef-c174-4cc0-82ff-ef3c459646fb)

user issue

Here is log when running LLM with meta-llama/Meta-Llama-3-8B-Instruct. Do you know how to train this model on a down-stream task? Thanks ----- Some weights of LlamaForCausalLM were not initialized from...

user issue

What are default values of max_generated_tokens, top_k, top_p, and temperature? If user doesn't set all parameters in `generate_kwargs` such as the example below, it should use default values. How do...

user issue

I try to transform a string into input llama2-specific and llama3-specific input in the function `completion_to_prompt()` Is there a way to pass parameter **model_option** as a input? or else, I...

user issue

I hope to switch llama2-7b-chat and llama3-8b models. But it cost a lot of memory size if I load both. How to clear one if I am going to load...

user issue

### Describe the bug To finetune model on Xeon CPU, we are following the [ai-reference-models/models_v2/pytorch/llama/training/cpu at main · intel/ai-reference-models (github.com)](https://github.com/intel/ai-reference-models/tree/main/models_v2/pytorch/llama/training/cpu) · Base model: meta-llama/llama-2-7b-hf · Dataset: [raw.githubusercontent.com/tloen/alpaca-lora/main/alpaca_data.json](https://raw.githubusercontent.com/tloen/alpaca-lora/main/alpaca_data.json) ### Versions ![image](https://github.com/user-attachments/assets/d58c72fb-807d-470d-84bc-c563c6b6daad)...

CPU
LLM
Escalate

**Describe the bug** sample https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/llm-agent-react/llm-agent-react.ipynb I converted a model for NPU. (Channel-wise quantization) `optimum-cli export openvino -m Qwen/Qwen2.5-3B-Instruct --weight-format int4 --sym --ratio 1.0 --group-size -1 Qwen2.5-3B-Instruct_NPU` Then, tried to run...

Stale

To compare multiple LLM models to OEM/ODM, we are running this sample. 1. install new env based on openvino guide: https://github.com/openvinotoolkit/openvino_notebooks/wiki/Windows 2. run >jupyter lab notebooks\llm-rag-llamaindex\llm-rag-llamaindex.ipynb 3. run the first...

support_request
Stale