We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Cuda:12.1 transformens:4.46.1 python:3.10 操作系统:windows 11 显卡:3090
V.0.16.2
1.pip install "xinference[transformers]" 2.xinference-local
1.pip install "xinference[transformers]" 2.xinference-local 3.浏览器打开http://localhost:9997 4.拉取glm4-chat,Engine选择transformers,quantization选择8-bit
如何解决这个问题:下载glm4-chat时,选择8-bit,下载报错:Server error: 500 - [address=127.0.0.1:4389, pid=15108] 'transfomg.word_embeddings.weight'
The text was updated successfully, but these errors were encountered:
This issue is stale because it has been open for 7 days with no activity.
Sorry, something went wrong.
No branches or pull requests
System Info / 系統信息
Cuda:12.1
transformens:4.46.1
python:3.10
操作系统:windows 11
显卡:3090
Running Xinference with Docker? / 是否使用 Docker 运行 Xinfernece?
Version info / 版本信息
V.0.16.2
The command used to start Xinference / 用以启动 xinference 的命令
1.pip install "xinference[transformers]"
2.xinference-local
Reproduction / 复现过程
1.pip install "xinference[transformers]"
2.xinference-local
3.浏览器打开http://localhost:9997
4.拉取glm4-chat,Engine选择transformers,quantization选择8-bit
Expected behavior / 期待表现
如何解决这个问题:下载glm4-chat时,选择8-bit,下载报错:Server error: 500 - [address=127.0.0.1:4389, pid=15108] 'transfomg.word_embeddings.weight'
The text was updated successfully, but these errors were encountered: