-
Notifications
You must be signed in to change notification settings - Fork 143
Issues: predibase/lorax
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Combining multiple LoRA adapters
question
Further information is requested
#429
opened Apr 21, 2024 by
winglian
LoRAX server with 2 GPUs and multiple adapters becomes permanently faster in swapping ONLY after parallel execution of requests.
question
Further information is requested
#395
opened Apr 8, 2024 by
lighteternal
1 of 4 tasks
Need some help. " You need to decrease --max-batch-prefill-tokens."
question
Further information is requested
#390
opened Apr 5, 2024 by
KrisWongz
4 tasks
Want Lorax with newer version of TGI
question
Further information is requested
#329
opened Mar 14, 2024 by
yangelaboy
performance issue
question
Further information is requested
#323
opened Mar 13, 2024 by
sleepwalker2017
Invoke API based external models
question
Further information is requested
#321
opened Mar 12, 2024 by
sekhar-hari
Sample command with mistral-7b failed
question
Further information is requested
#312
opened Mar 7, 2024 by
hayleyhu
2 of 4 tasks
decapoda-research/llama-13b-hf is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models'
question
Further information is requested
#310
opened Mar 7, 2024 by
sleepwalker2017
Error while running the pre-built container using Podman
question
Further information is requested
#266
opened Feb 21, 2024 by
chaser06
Can the basemodel be loaded directly from the local system without connecting to Hugging Face? It cannot be launched because there is no connection.
question
Further information is requested
#261
opened Feb 20, 2024 by
Loovelj
2 tasks
Multiple base models
question
Further information is requested
#253
opened Feb 16, 2024 by
sekhar-hari
1 of 2 tasks
Server error: This model was initialized with the adapter xxx and therefore does not support dynamic adapter loading. Please initialize a new model instance from the base model in order to use the dynamic adapter loading feature.
question
Further information is requested
#244
opened Feb 14, 2024 by
avoskresensky
2 of 4 tasks
Mixtral nf4 performance 2x slower than expected
question
Further information is requested
#211
opened Jan 29, 2024 by
timohear
2 of 4 tasks
Support self-trained model
question
Further information is requested
#208
opened Jan 25, 2024 by
yangelaboy
2 tasks
ValueError: Adapter '/data/llama2-lora' is not compatible with model '/data/Llama-2-7b-chat-hf'. Use --model-id '/new-model/llama2-7b/Llama-2-7b-chat-hf' instead.
question
Further information is requested
#172
opened Jan 10, 2024 by
Senna1960321
2 of 4 tasks
Second GPU is not found when running --sharded true
question
Further information is requested
#150
opened Dec 24, 2023 by
psych0v0yager
2 of 4 tasks
Latency increase when run on multi-GPU
question
Further information is requested
#116
opened Dec 8, 2023 by
prd-tuong-nguyen
2 of 4 tasks
Some error records and questions
question
Further information is requested
#115
opened Dec 8, 2023 by
KrisWongz
1 of 4 tasks
Question regarding Punica integeration
question
Further information is requested
#107
opened Dec 6, 2023 by
psych0v0yager
How to use --master-addr <MASTER_ADDR>|--master-port <MASTER_PORT>?
question
Further information is requested
#99
opened Dec 4, 2023 by
prd-tuong-nguyen
2 of 4 tasks
how does this differ from s-Lora?
question
Further information is requested
#90
opened Nov 30, 2023 by
priyankat99
ProTip!
Follow long discussions with comments:>50.