Skip to content

Commit 721361f

Browse files
authored
DOC: Update xllamacpp doc (#3862)
1 parent 528b332 commit 721361f

File tree

5 files changed

+147
-173
lines changed

5 files changed

+147
-173
lines changed
107 KB
Loading

doc/source/getting_started/installation.rst

Lines changed: 2 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -125,22 +125,9 @@ and is the sole backend for llama.cpp since v1.6.0.
125125

126126
Initial setup::
127127

128-
pip install xinference
129-
130-
Installation instructions for ``xllamacpp``:
131-
132-
- CPU or Mac Metal::
133-
134-
pip install -U xllamacpp
135-
136-
- CUDA::
137-
138-
pip install xllamacpp --force-reinstall --index-url https://xorbitsai.github.io/xllamacpp/whl/cu124
139-
140-
- HIP::
141-
142-
pip install xllamacpp --force-reinstall --index-url https://xorbitsai.github.io/xllamacpp/whl/rocm-6.0.2
128+
pip install "xinference[llama_cpp]"
143129

130+
For more installation instructions for ``xllamacpp`` to enable GPU acceleration, please refer to: https://github.com/xorbitsai/xllamacpp
144131

145132
SGLang Backend
146133
~~~~~~~~~~~~~~

doc/source/locale/zh_CN/LC_MESSAGES/getting_started/installation.po

Lines changed: 17 additions & 25 deletions
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ msgid ""
77
msgstr ""
88
"Project-Id-Version: Xinference \n"
99
"Report-Msgid-Bugs-To: \n"
10-
"POT-Creation-Date: 2025-07-29 21:39+0800\n"
10+
"POT-Creation-Date: 2025-07-30 11:01+0800\n"
1111
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
1212
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
1313
"Language: zh_CN\n"
@@ -16,7 +16,7 @@ msgstr ""
1616
"MIME-Version: 1.0\n"
1717
"Content-Type: text/plain; charset=utf-8\n"
1818
"Content-Transfer-Encoding: 8bit\n"
19-
"Generated-By: Babel 2.14.0\n"
19+
"Generated-By: Babel 2.17.0\n"
2020

2121
#: ../../source/getting_started/installation.rst:5
2222
msgid "Installation"
@@ -44,8 +44,8 @@ msgid ""
4444
"sglang, please install it separately via ``pip install "
4545
"'xinference[sglang]'``."
4646
msgstr ""
47-
"由于 vllm 和 sglang 在包依赖上无法调和,因此,我们从 all 里移除了 sglang"
48-
"如果要使用 sglang,请使用 ``pip install 'xinference[sglang]'`` 。"
47+
"由于 vllm 和 sglang 在包依赖上无法调和,因此,我们从 all 里移除了 sglang"
48+
"如果要使用 sglang,请使用 ``pip install 'xinference[sglang]'`` 。"
4949

5050
#: ../../source/getting_started/installation.rst:17
5151
msgid "Several usage scenarios require special attention."
@@ -347,32 +347,24 @@ msgstr ""
347347
"v1.6.0 开始,该后端已被移除。"
348348

349349
#: ../../source/getting_started/installation.rst:125
350-
#: ../../source/getting_started/installation.rst:148
351-
#: ../../source/getting_started/installation.rst:161
350+
#: ../../source/getting_started/installation.rst:135
351+
#: ../../source/getting_started/installation.rst:144
352352
msgid "Initial setup::"
353353
msgstr "初始步骤:"
354354

355355
#: ../../source/getting_started/installation.rst:129
356-
msgid "Installation instructions for ``xllamacpp``:"
357-
msgstr "``xllamacpp`` 的安装说明:"
358-
359-
#: ../../source/getting_started/installation.rst:131
360-
msgid "CPU or Mac Metal::"
361-
msgstr "CPU 或 Mac Metal:"
362-
363-
#: ../../source/getting_started/installation.rst:135
364-
msgid "CUDA::"
365-
msgstr ""
366-
367-
#: ../../source/getting_started/installation.rst:139
368-
msgid "HIP::"
356+
msgid ""
357+
"For more installation instructions for ``xllamacpp`` to enable GPU "
358+
"acceleration, please refer to: https://github.com/xorbitsai/xllamacpp"
369359
msgstr ""
360+
"更多的 ``xllamacpp`` 安装说明以便开启 GPU 加速,请参考:https://github.com"
361+
"/xorbitsai/xllamacpp"
370362

371-
#: ../../source/getting_started/installation.rst:145
363+
#: ../../source/getting_started/installation.rst:132
372364
msgid "SGLang Backend"
373365
msgstr "SGLang 引擎"
374366

375-
#: ../../source/getting_started/installation.rst:146
367+
#: ../../source/getting_started/installation.rst:133
376368
msgid ""
377369
"SGLang has a high-performance inference runtime with RadixAttention. It "
378370
"significantly accelerates the execution of complex LLM programs by "
@@ -383,18 +375,18 @@ msgstr ""
383375
"自动重用KV缓存,显著加速了复杂 LLM 程序的执行。它还支持其他常见推理技术,"
384376
"如连续批处理和张量并行处理。"
385377

386-
#: ../../source/getting_started/installation.rst:158
378+
#: ../../source/getting_started/installation.rst:141
387379
msgid "MLX Backend"
388380
msgstr "MLX 引擎"
389381

390-
#: ../../source/getting_started/installation.rst:159
382+
#: ../../source/getting_started/installation.rst:142
391383
msgid "MLX-lm is designed for Apple silicon users to run LLM efficiently."
392384
msgstr "MLX-lm 用来在苹果 silicon 芯片上提供高效的 LLM 推理。"
393385

394-
#: ../../source/getting_started/installation.rst:166
386+
#: ../../source/getting_started/installation.rst:149
395387
msgid "Other Platforms"
396388
msgstr "其他平台"
397389

398-
#: ../../source/getting_started/installation.rst:168
390+
#: ../../source/getting_started/installation.rst:151
399391
msgid ":ref:`Ascend NPU <installation_npu>`"
400392
msgstr ""

0 commit comments

Comments
 (0)