A high-throughput and memory-efficient inference and serving engine for LLMs
-
Updated
Apr 23, 2025 - Python
A high-throughput and memory-efficient inference and serving engine for LLMs
🎨ComfyUI standalone pack for Intel GPUs. | 英特尔显卡 ComfyUI 整合包
Add a description, image, and links to the xpu topic page so that developers can more easily learn about it.
To associate your repository with the xpu topic, visit your repo's landing page and select "manage topics."