-
Notifications
You must be signed in to change notification settings - Fork 1.1k
Issues: abetlen/llama-cpp-python
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
using Further information is requested
cmake .. -DBUILD_SHARED_LIBS=ON
failed
question
#1321
opened Apr 3, 2024 by
LUXUS1
why is it so slower than origin llava1.5 model?
question
Further information is requested
#1313
opened Mar 29, 2024 by
ganliqiang
llama-cpp-python bindings not working for multiple GPUs
question
Further information is requested
#1310
opened Mar 28, 2024 by
y6t4
Retrieve attention score for all input tokens per generated token
enhancement
New feature or request
question
Further information is requested
#1141
opened Jan 29, 2024 by
parallaxe
Anyone built with vulkan yet?
question
Further information is requested
#1138
opened Jan 28, 2024 by
Ph0rk0z
Smaller docker image CUDA example
question
Further information is requested
#1119
opened Jan 23, 2024 by
Hansson0728
Concurrent request handling
question
Further information is requested
#1062
opened Jan 4, 2024 by
khanjandharaiya
LlamaCPP Usage
question
Further information is requested
#1035
opened Dec 21, 2023 by
4entertainment
Include in Readme how to Pass Custom Arguments to Improvements or additions to documentation
question
Further information is requested
llama_cpp.server
in Docker
documentation
#1029
opened Dec 19, 2023 by
jaredquekjz
Support for a limited vocabulary for generation
question
Further information is requested
#998
opened Dec 11, 2023 by
mgorenstein
I want to know how to enable AMD GPU or enable hipBLAS/ROCm on Windows. I tried the official HIP (sdk) and set the relevant environment flags, but it seems to have no effect. BLAS is always at 0. Is there any solution for Windows? Plan?
question
Further information is requested
#978
opened Dec 6, 2023 by
chengjia604
How can I extract data from documents in JSON Output Format?
question
Further information is requested
#942
opened Nov 24, 2023 by
garciacarmonaam
How to add data to a llama2tian model that was not previously in the model through langchain via port 8000.
question
Further information is requested
#933
opened Nov 22, 2023 by
AiTaiShan
Help needed with loading 'TheBloke/Mistral-7B-Instruct-v0.1-GGUF' model using Further information is requested
llama-cpp-python
question
#915
opened Nov 15, 2023 by
ElliotBadinger
Intel Arc Support?
enhancement
New feature or request
question
Further information is requested
#901
opened Nov 11, 2023 by
ElliottDyson
Llama model entering into a lenghty question answer mode
question
Further information is requested
#885
opened Nov 7, 2023 by
AvisP
Why can't multiple apis be triggered at the same time
question
Further information is requested
#873
opened Nov 5, 2023 by
zhengzhanpeng
Second GPU detected by "nvidia-smi", but not detected when running CodeLlama with Llama-cpp-python
question
Further information is requested
#754
opened Sep 26, 2023 by
mersu898
[Q] Is the prompt/chat history stored on disk?
question
Further information is requested
#737
opened Sep 19, 2023 by
NightMachinery
I have some params in llama.cpp but i cant see in this repo
question
Further information is requested
#731
opened Sep 18, 2023 by
illumionous
model_path error with Llama-2
question
Further information is requested
#526
opened Jul 25, 2023 by
jadehardouin
Count/truncate number of tokens before processing
enhancement
New feature or request
question
Further information is requested
#103
opened Apr 22, 2023 by
jakvb
ProTip!
Follow long discussions with comments:>50.