Wang, Kai Lawrence

Results 15 comments of Wang, Kai Lawrence

> How can I try this fix? You can apply the change of `CodeGen/codegen.py` in the PR to your code, rebuild the megaservice docker image (`opea/codegen:latest`) then restart the services...

I've verified that this PR fixed the output format issue originally, the accuracy benchmark could run smoothly to the end for `Qwen/Qwen2.5-Coder-7B-Instruct` and `Qwen/CodeQwen1.5-7B-Chat`. But there are still some issues...

@YYue000 Thanks for reporting this issue. We will look into it and give feedback later.

@YYue000 I can reproduce this issue. Using `2.4.0+cpu`, a) Without `ipex.llm.optimize()`, invoking `model.generate()` with `return_dict_in_generate=True` would return an object type of `transformers.generation.utils.GenerateDecoderOnlyOutput` where attributes like `sequences` can be retrieved. b)...

> @YYue000 I can reproduce this issue. Using `2.4.0+cpu`, a) Without `ipex.llm.optimize()`, invoking `model.generate()` with `return_dict_in_generate=True` would return an object type of `transformers.generation.utils.GenerateDecoderOnlyOutput` where attributes like `sequences` can be retrieved....