Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Question]: Does text-embedding-v2 use a local model? #1298

Open
leecj opened this issue Jun 27, 2024 · 3 comments
Open

[Question]: Does text-embedding-v2 use a local model? #1298

leecj opened this issue Jun 27, 2024 · 3 comments
Labels
question Further information is requested

Comments

@leecj
Copy link

leecj commented Jun 27, 2024

Describe your problem

我使用了通义千问的API key,embedding模型选择text-embedding-v2,为什么CPU模式下embedding非常缓慢,并且后台在下载text-embedding-v2模型

I used the API key from Tongyi Qianwen, and selected the embedding model text-embedding-v2. Why is the embedding very slow in CPU mode, and is the backend downloading the text-embedding-v2 model?

image
@leecj leecj added the question Further information is requested label Jun 27, 2024
@yangboz
Copy link

yangboz commented Jun 27, 2024

Describe your problem

我使用了通义千问的API key,embedding模型选择text-embedding-v2,为什么CPU模式下embedding非常缓慢,并且后台在下载text-embedding-v2模型

I used the API key from Tongyi Qianwen, and selected the embedding model text-embedding-v2. Why is the embedding very slow in CPU mode, and is the backend downloading the text-embedding-v2 model?

image

可能需要预先手动执行如下命令:
export HF_ENDPOINT=https://hf-mirror.com
然后下载模型至本地缓存
huggingface-cli download --resume-download text-embedding-v2 --local-dir text-embedding-v2

@KevinHuSh
Copy link
Collaborator

Pull the dev version of docker image and git pull the latest code.

@leecj
Copy link
Author

leecj commented Jun 28, 2024

Pull the dev version of docker image and git pull the latest code.

我还是不懂为什么需要本地的text-embedding-v2。另外当聊天时下载模型的设计非常失败,这不利于软件设计的稳定性,建议在启动、配置时就保障模型已经下载完成

I still don't understand why we need the local text-embedding-v2. Additionally, the design of downloading the model during the chat is very poor, which is not conducive to the stability of the software design. It is recommended that the model should be downloaded and completed during startup and configuration.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

3 participants