We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
chatglm2-6b-int4的模型大概在4G左右,我看LinkSoul/Chinese-Llama-2-7b-4bit的模型大小在13G左右
The text was updated successfully, but these errors were encountered:
量化方法不同 可以考慮用llamacpp能運行的ggml版本
Sorry, something went wrong.
请问本地加载的时候如何分布式加载呢,有多个gpu会自动做分片加载吗,单gpu装不下
No branches or pull requests
chatglm2-6b-int4的模型大概在4G左右,我看LinkSoul/Chinese-Llama-2-7b-4bit的模型大小在13G左右
The text was updated successfully, but these errors were encountered: