Skip to content

Issues: abetlen/llama-cpp-python

Roadmap for v0.2
#487 opened Jul 18, 2023 by abetlen
Open 1
Add batched inference
#771 opened Sep 30, 2023 by abetlen
Open 37
Improve installation process
#1178 opened Feb 12, 2024 by abetlen
Open 8
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Issues list

Feat: Add support for Llama 3.1 function calling enhancement New feature or request
#1618 opened Jul 24, 2024 by qnixsynapse
Add support for croos-encoders enhancement New feature or request
#1611 opened Jul 20, 2024 by perpendicularai
Pull from Ollama repo functionality enhancement New feature or request
#1607 opened Jul 18, 2024 by ericcurtin
How to log raw token generation? enhancement New feature or request
#1546 opened Jun 21, 2024 by sisi399
Multi-arch support for pre-built cpu wheel enhancement New feature or request
#1506 opened Jun 5, 2024 by abetlen
Please add response_format to create_completion enhancement New feature or request
#1478 opened May 23, 2024 by dtkettler
arm64 builds for CUDA enhancement New feature or request
#1446 opened May 10, 2024 by mcvella
Add Nous Hermes 2 Pro function calling ChatHandler. enhancement New feature or request
#1429 opened May 5, 2024 by stygmate
Is loading of control vectors supported? enhancement New feature or request
#1363 opened Apr 19, 2024 by edwinRNDR
[REQUEST] Accept raw token IDs in stop parameter enhancement New feature or request
#1360 opened Apr 18, 2024 by ddh0
Hermes 2 Pro Full Chat Format Support enhancement New feature or request
#1339 opened Apr 9, 2024 by abetlen
Allow any format for X-Request-Id enhancement New feature or request
#1337 opened Apr 9, 2024 by ging-dev
4 tasks done
Models with multiple chat templates enhancement New feature or request
#1336 opened Apr 8, 2024 by CISC
Add command-r support like llamacpp has enhancement New feature or request
#1279 opened Mar 16, 2024 by rombodawg
Does this lib support contrastive search decoding ? enhancement New feature or request
#1253 opened Mar 5, 2024 by congson1293
Add Self-Extend support? enhancement New feature or request
#1242 opened Mar 1, 2024 by theaerotoad
llama_cpp.server save chat log enhancement New feature or request
#1224 opened Feb 26, 2024 by riverzhou
Improve installation process enhancement New feature or request help wanted Extra attention is needed
#1178 opened Feb 12, 2024 by abetlen
ProTip! What’s not been updated in a month: updated:<2025-04-25.