You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
feat: save and restore a context sequence state (#460)
* feat: save and restore a context sequence state
* feat: stream function call parameters
* feat: configure Hugging Face remote endpoint for resolving URIs
* feat: Qwen 3 support
* feat(`QwenChatWrapper`): support discouraging the generation of thoughts
* feat(`getLlama`): `dryRun` option
* feat: `getLlamaGpuTypes` function
* fix: adapt to breaking `llama.cpp` changes
* fix: capture multi-token segment separators
* fix: race condition when reading extremely long gguf metadata
* fix: adapt memory estimation to new added model architectures
* fix: skip binary testing on certain problematic conditions
* fix: improve GPU backend loading error description
* fix: update gguf types
* fix: performance improvements
* docs: update the awesome list
* docs: solutions to more CUDA issues
Copy file name to clipboardExpand all lines: docs/guide/CUDA.md
+27Lines changed: 27 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -114,6 +114,33 @@ set NODE_LLAMA_CPP_CMAKE_OPTION_CMAKE_GENERATOR_TOOLSET=%CUDA_PATH%
114
114
115
115
Then run the build command again to check whether setting the `CMAKE_GENERATOR_TOOLSET` cmake option fixed the issue.
116
116
117
+
### Fix the `forward compatibility was attempted on non supported HW` Error {#fix-cuda-forward-compatibility}
118
+
This error usually happens when the CUDA version you have installed on your machine is older than the CUDA version used in the prebuilt binaries supplied by `node-llama-cpp`.
119
+
120
+
To resolve this issue, you can either [update your CUDA installation](https://developer.nvidia.com/cuda-downloads) to the latest version (recommended) or [build `node-llama-cpp` on your machine](#building) against the CUDA version you have installed.
121
+
122
+
### Fix the `Binary GPU type mismatch. Expected: cuda, got: false` Error {#fix-cuda-gpu-type-mismatch}
123
+
This error usually happens when you have multiple conflicting CUDA versions installed on your machine.
124
+
125
+
To fix it, uninstall older CUDA versions and restart your machine (important).
126
+
127
+
:::: details Check which CUDA libraries are picked up by `node-llama-cpp`'s prebuilt binaries on your machine
0 commit comments