[Question]: Does text-embedding-v2 use a local model?
Describe your problem
我使用了通义千问的API key,embedding模型选择text-embedding-v2,为什么CPU模式下embedding非常缓慢,并且后台在下载text-embedding-v2模型
I used the API key from Tongyi Qianwen, and selected the embedding model text-embedding-v2. Why is the embedding very slow in CPU mode, and is the backend downloading the text-embedding-v2 model?
Describe your problem
我使用了通义千问的API key,embedding模型选择text-embedding-v2,为什么CPU模式下embedding非常缓慢,并且后台在下载text-embedding-v2模型
I used the API key from Tongyi Qianwen, and selected the embedding model text-embedding-v2. Why is the embedding very slow in CPU mode, and is the backend downloading the text-embedding-v2 model?
![]()
可能需要预先手动执行如下命令: export HF_ENDPOINT=https://hf-mirror.com 然后下载模型至本地缓存 huggingface-cli download --resume-download text-embedding-v2 --local-dir text-embedding-v2
Pull the dev version of docker image and git pull the latest code.
Pull the dev version of docker image and git pull the latest code.
我还是不懂为什么需要本地的text-embedding-v2。另外当聊天时下载模型的设计非常失败,这不利于软件设计的稳定性,建议在启动、配置时就保障模型已经下载完成
I still don't understand why we need the local text-embedding-v2. Additionally, the design of downloading the model during the chat is very poor, which is not conducive to the stability of the software design. It is recommended that the model should be downloaded and completed during startup and configuration.