-
Notifications
You must be signed in to change notification settings - Fork 1.3k
Issues: LlamaFamily/Llama-Chinese
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
AttributeError: type object 'AttentionMaskConverter' has no attribute '_ignore_causal_mask_sdpa'
#333
opened May 3, 2024 by
gongye19
这个需要什么配置合适?用一张A100 显卡跑的7B模型,80G显存用了10G,回答case中的怎么去北京 要60秒才返回结果
#310
opened Apr 3, 2024 by
MetaRunning
RuntimeError: FlashAttention only supports Ampere GPUs or newer.
#307
opened Mar 21, 2024 by
540627735
ProTip!
Updated in the last three days: updated:>2024-12-23.