ppl.pmx
ppl.pmx copied to clipboard
`\{softmax}` is not supported in KaTeX, a change to `softmax` can fix it. 
Hi guys, I see your project in OSPP. I'm wondering are there any good issues for kick start?
使用的model_zoo/qwen/huggingface/ConvertWeightToOpmx.py,py文件内param参数全都是对应qwen而非qwen1.5设置的。无法找到qwen1.5权重转换对应的脚本。
add llama pipeline parallel, do not need to do extra split model. It split tp and pp while loading model
Support int4 weight only quantization for Llama3 1. Define the weight only layer in ModelParallel.py 2. Define ConvertWeightToOpmx.py and add quant here 3. Update Dynamic Static modeling for quant model...
add setup.py for opmx