-
Notifications
You must be signed in to change notification settings - Fork 4
Home
Redefining how AI APIs communicate with the web — built from scratch in C and Assembly.
NeuroHTTP is a next-generation web server built entirely in C and Assembly, designed specifically for AI workloads.
While traditional servers like NGINX or Apache were built for static content or general-purpose APIs, NeuroHTTP is engineered for:
- 🧠 AI streaming responses (LLMs, chatbots, real-time inference)
- ⚡ Massive concurrent API requests
- 📦 Low-latency JSON handling
- 🔌 Real-time communication (HTTP/3, WebSocket, gRPC)
Goal: to create the world’s first AI-native web server — optimized from the kernel upward for intelligent workloads.
“The web was built for documents. Then came applications. Now it’s time for AI.”
NeuroHTTP aims to redefine how AI models are served at scale, introducing a native AI transport layer that’s:
- Fast ⚡
- Modular 🧩
- Open Source 🌍
| Feature | Description |
|---|---|
| ⚙️ AI Stream Mode | Token-by-token streaming for LLMs and inference APIs |
| 🧠 AI Router | Dynamic routing between multiple models (GPT, LLaMA, Claude…) |
| 🧰 Plugin System | Extendable via runtime C modules — no recompilation needed |
| 🧮 Assembly-Optimized Core | Low-level JSON and I/O parsing for extreme speed |
| 🔐 Built-in Security | Token validation, request quotas, and internal firewall |
| 📊 Telemetry System | Real-time metrics: latency, throughput, memory usage |
NeuroHTTP follows a modular, event-driven architecture, built for performance and extensibility:
- Core Engine: C-based HTTP and TCP stack with epoll and threading
- Worker Threads: Adaptive scheduling per workload
- AI Router: Direct model inference routing and optimization
- Plugin Layer: C modules for custom AI adapters or routing logic
- Cache System: TTL-based memory cache for reusable responses
Every subsystem is built from scratch — no dependency on NGINX, Node, or Go — ensuring full control and efficiency.
| Traditional Servers | NeuroHTTP |
|---|---|
| Optimized for static content | Optimized for AI inference |
| Blocking or event-loop concurrency | Adaptive, multi-threaded concurrency |
| JSON parsing at user-level | Assembly-level parsing |
| Separate model APIs | Built-in AI model router |
| No real AI streaming | Native AI Stream Mode |
git clone https://github.com/okba14/NeuroHTTP.git
cd NeuroHTTP
make all
./bin/aionic
Access the default endpoint:
curl http://localhost:8080
🌍 Open Source Goals
NeuroHTTP will be released under the MIT License, encouraging:
Collaboration from the open-source community
Benchmark comparisons with NGINX, Envoy, and Caddy
Integration into AI backends and frameworks like LangChain
🧑💻 Author
👨💻 GUIAR OQBA — Creator of NeuroHTTP
“Empowering the next generation of AI-native infrastructure — from Algeria 🇩🇿.”
GitHub • LinkedIn • Discussions
Fast. Modular. AI-Native. That’s NeuroHTTP.
Join the mission to redefine how the web talks to AI — one packet at a time.
⚡ NeuroHTTP — An Open-Source AI Web Server
Built with ❤️ by GUIAR OQBA & contributors.
📚 Documentation •
💬 Join the Discussion •
🐞 Report a Bug
MIT Licensed • Algeria 🇩🇿 • 2025