Skip to content
#

openai-compatible

Here are 48 public repositories matching this topic...

⚡ Python-free Rust inference server — OpenAI-API compatible. GGUF + SafeTensors, hot model swap, auto-discovery, single binary. FREE now, FREE forever.

  • Updated Oct 23, 2025
  • Rust

A high-performance API server that provides OpenAI-compatible endpoints for MLX models. Developed using Python and powered by the FastAPI framework, it provides an efficient, scalable, and user-friendly solution for running MLX-based vision and language models locally with an OpenAI-compatible interface.

  • Updated Nov 24, 2025
  • Python

All-in-one local AI stack for Mistral-13B and Llama.cpp, with one-step CUDA wheel install, OpenAI-compatible API, and modern web dashboard. Switch between local and cloud chat, run on your own GPU, and deploy instantly—no API keys or paywalls. Designed for easy install, custom builds, and fast remote access. Enjoy!

  • Updated Aug 16, 2025
  • Python

A FastAPI gateway for local LLMs that adds intelligent web research, multilingual recency/how-to detection, time-anchored guidance, context injection, and OpenAI-compatible SSE streaming. Turn any local model into a recency-aware, context-enhanced assistant instantly.

  • Updated Nov 20, 2025
  • Python

Improve this page

Add a description, image, and links to the openai-compatible topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the openai-compatible topic, visit your repo's landing page and select "manage topics."

Learn more