Popular repositories Loading
-
exllamav2
exllamav2 PublicForked from turboderp-org/exllamav2
A fast inference library for running LLMs locally on modern consumer-class GPUs
Python
-
tabbyAPI
tabbyAPI PublicForked from theroyallab/tabbyAPI
An OAI compatible exllamav2 API that's both lightweight and fast
Python
-
runpod-worker-tabbyapi
runpod-worker-tabbyapi Public templateForked from runpod-workers/worker-template
Python
-
worker-vllm
worker-vllm PublicForked from runpod-workers/worker-vllm
The RunPod worker template for serving our large language model endpoints. Powered by vLLM.
Python
If the problem persists, check the GitHub status page or contact support.