garycao
garycao
@simon-mo can you help us review the code?
> Getting > > ``` > Traceback (most recent call last): > File "/opt/conda/lib/python3.10/runpy.py", line 196, in _run_module_as_main > return _run_code(code, main_globals, None, > File "/opt/conda/lib/python3.10/runpy.py", line 86, in _run_code...
> Getting > > ``` > Traceback (most recent call last): > File "/opt/conda/lib/python3.10/runpy.py", line 196, in _run_module_as_main > return _run_code(code, main_globals, None, > File "/opt/conda/lib/python3.10/runpy.py", line 86, in _run_code...
> I'm running into the following issues: > > * Completion not working > * Chat template is missing in tokenizer config, the default one will just keep the generation...
> The tokenizer_config.json should also include the following so it doesn't need to be specified by the client each time. Please let me know once the hf or modelscope version...
> The tokenizer_config.json should also include the following so it doesn't need to be specified by the client each time. Please let me know once the hf or modelscope version...
> ./format.sh OK ``` $ ./format.sh vLLM yapf: Done vLLM mypy: Success: no issues found in 3 source files Success: no issues found in 7 source files Success: no issues...
> Also feel free to add to https://github.com/vllm-project/vllm/blob/main/docs/source/models/supported_models.rst and README https://github.com/vllm-project/vllm?tab=readme-ov-file#about updated. thanks
merged v0.4.2 [[Misc][Refactor] Generalize linear_method to be quant_method (https://github.com/vllm-project/vllm/pull/4373)](https://github.com/vllm-project/vllm/commit/a62aaf1df558d69658a42c1ab749368ab0325f35)
@simon-mo is it ready to merge?