Skip to content

Issues: abetlen/llama-cpp-python

Roadmap for v0.2
#487 opened Jul 18, 2023 by abetlen
Open 1
Add batched inference
#771 opened Sep 30, 2023 by abetlen
Open 37
Improve installation process
#1178 opened Feb 12, 2024 by abetlen
Open 8
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Issues list

igpu question Further information is requested
#1709 opened Aug 26, 2024 by ayttop
get hidden state question Further information is requested
#1695 opened Aug 21, 2024 by fousdfrf
using cmake .. -DBUILD_SHARED_LIBS=ON failed question Further information is requested
#1321 opened Apr 3, 2024 by LUXUS1
why is it so slower than origin llava1.5 model? question Further information is requested
#1313 opened Mar 29, 2024 by ganliqiang
llama-cpp-python bindings not working for multiple GPUs question Further information is requested
#1310 opened Mar 28, 2024 by y6t4
Retrieve attention score for all input tokens per generated token enhancement New feature or request question Further information is requested
#1141 opened Jan 29, 2024 by parallaxe
Anyone built with vulkan yet? question Further information is requested
#1138 opened Jan 28, 2024 by Ph0rk0z
Smaller docker image CUDA example question Further information is requested
#1119 opened Jan 23, 2024 by Hansson0728
Concurrent request handling question Further information is requested
#1062 opened Jan 4, 2024 by khanjandharaiya
LlamaCPP Usage question Further information is requested
#1035 opened Dec 21, 2023 by 4entertainment
Include in Readme how to Pass Custom Arguments to llama_cpp.server in Docker documentation Improvements or additions to documentation question Further information is requested
#1029 opened Dec 19, 2023 by jaredquekjz
Support for a limited vocabulary for generation question Further information is requested
#998 opened Dec 11, 2023 by mgorenstein
How can I extract data from documents in JSON Output Format? question Further information is requested
#942 opened Nov 24, 2023 by garciacarmonaam
Intel Arc Support? enhancement New feature or request question Further information is requested
#901 opened Nov 11, 2023 by ElliottDyson
Llama model entering into a lenghty question answer mode question Further information is requested
#885 opened Nov 7, 2023 by AvisP
Why can't multiple apis be triggered at the same time question Further information is requested
#873 opened Nov 5, 2023 by zhengzhanpeng
[Q] Is the prompt/chat history stored on disk? question Further information is requested
#737 opened Sep 19, 2023 by NightMachinery
I have some params in llama.cpp but i cant see in this repo question Further information is requested
#731 opened Sep 18, 2023 by illumionous
Arabic language question Further information is requested
#665 opened Sep 4, 2023 by mitha469
model_path error with Llama-2 question Further information is requested
#526 opened Jul 25, 2023 by jadehardouin
Count/truncate number of tokens before processing enhancement New feature or request question Further information is requested
#103 opened Apr 22, 2023 by jakvb
ProTip! Follow long discussions with comments:>50.