LongBench
LongBench copied to clipboard
怎么测试qwen2.5系列模型
你好,qwen2.5系列模型的评估是需要更改config的max_position_embedding为120000吗?
需要用YaRN,以下是qwen官方给的部署教程:
The current config.json is set for context length up to 32,768 tokens. To handle extensive inputs exceeding 32,768 tokens, we utilize YaRN, a technique for enhancing model length extrapolation, ensuring optimal performance on lengthy texts.
For supported frameworks, you could add the following to config.json to enable YaRN:
{
...,
"rope_scaling": {
"factor": 4.0,
"original_max_position_embeddings": 32768,
"type": "yarn"
}
}
论文给出了使用或不使用 YaRN 的对比:
但是设置了YARN还是被报超长的问题