-
Notifications
You must be signed in to change notification settings - Fork 12.7k
Comparing changes
Open a pull request
base repository: ggml-org/llama.cpp
base: master@{1day}
head repository: ggml-org/llama.cpp
compare: master
- 16 commits
- 46 files changed
- 11 contributors
Commits on Aug 14, 2025
-
Configuration menu - View commit details
-
Copy full SHA for 8b24837 - Browse repository at this point
Copy the full SHA 8b24837View commit details -
ggml: fix ggml_conv_1d_dw bug (ggml/1323)
* ggml: fix ggml_conv_1d_dw bug * Fixed conv1d_dw weight tensor dimension.
Configuration menu - View commit details
-
Copy full SHA for 5ade300 - Browse repository at this point
Copy the full SHA 5ade300View commit details -
Configuration menu - View commit details
-
Copy full SHA for 3973163 - Browse repository at this point
Copy the full SHA 3973163View commit details -
server : add SWA checkpoints (#15293)
* server : add SWA checkpoints ggml-ci * cont : server clean-up * server : handle state restore fails * llama : add extended llama_state_seq_ API * server : do not make checkpoints if --swa-full ggml-ci * llama : remove flags value for NONE * server : configure number of SWA checkpoints with CLI arg ggml-ci * args : fix scope of new argument
Configuration menu - View commit details
-
Copy full SHA for d32e03f - Browse repository at this point
Copy the full SHA d32e03fView commit details -
vulkan: perf_logger improvements (#15246)
* vulkan: perf_logger improvements - Account for batch dimension in flops calculation. - Fix how "_VEC" is detected for mat_mul_id. - Fix "n" dimension for mat_mul_id (in case of broadcasting). - Include a->type in name. * use <=mul_mat_vec_max_cols rather than ==1
Configuration menu - View commit details
-
Copy full SHA for 863d341 - Browse repository at this point
Copy the full SHA 863d341View commit details -
Configuration menu - View commit details
-
Copy full SHA for 1a01899 - Browse repository at this point
Copy the full SHA 1a01899View commit details -
Configuration menu - View commit details
-
Copy full SHA for 646944c - Browse repository at this point
Copy the full SHA 646944cView commit details -
gpt-oss: implement harmony parsing (#15181)
* model : add harmony parser for gpt-oss * gpt-oss : fix grammar trigger from causing empty stack * gpt-oss: tweak the grammar trigger again * gpt-oss : add support for recipient in role header * gpt-oss : fix ungrouped tool calls in grammar * gpt-oss : loosen function name matching during parse * gpt-oss : clean up workarounds * gpt-oss : add template tests * gpt-oss : simulate thinking and tool call tags * gpt-oss : undo think tags when reasoning_format is none * gpt-oss : set special tokens back to user defined * gpt-oss : update openai-gpt-oss template * server : filter out harmony thought messages * gpt-oss : simplify parsing
Configuration menu - View commit details
-
Copy full SHA for b204a5a - Browse repository at this point
Copy the full SHA b204a5aView commit details -
HIP: Cleanup hipification header (#15285)
add expicit conversion operator to support older versions of rocm Switch over to hip_bf16 from legacy hip_bfloat16 Simplify RDNA3 define Reduce swap over of new hipblas api to rocm 6.5 as this version is used for rocm 7.0 previews --------- Co-authored-by: Johannes Gäßler <johannesg@5d6.de>
Configuration menu - View commit details
-
Copy full SHA for 5ba36f6 - Browse repository at this point
Copy the full SHA 5ba36f6View commit details -
Configuration menu - View commit details
-
Copy full SHA for e4e9159 - Browse repository at this point
Copy the full SHA e4e9159View commit details -
llama : add 18-layer model type for Gemma 3-270m (#15319)
This commit adds support for the 18-layer model type in the Gemma3 series, which is the size of the Gemma3-270m model. The motivation for this commit is was the only change required for Gemma3-270m to be converted to GGUF format and used with llama.cpp. Once the model has been converted and uploaded to Huggingface it can be used like this: ```console $ ./build/bin/llama-cli -hf ggml-org/gemma-3-270m-GGUF:Q8_0 ```
Configuration menu - View commit details
-
Copy full SHA for 7a0de96 - Browse repository at this point
Copy the full SHA 7a0de96View commit details -
Configuration menu - View commit details
-
Copy full SHA for f75b830 - Browse repository at this point
Copy the full SHA f75b830View commit details -
eval-callback : stop on first NaN (#15320)
* eval-callback : stop on first NaN * cont : log error
Configuration menu - View commit details
-
Copy full SHA for df36bce - Browse repository at this point
Copy the full SHA df36bceView commit details -
Configuration menu - View commit details
-
Copy full SHA for 4227c9b - Browse repository at this point
Copy the full SHA 4227c9bView commit details
Commits on Aug 15, 2025
-
test-opt: fix backend support check (#15317)
* test-opt: fix backend support check * Update tests/test-opt.cpp Co-authored-by: Georgi Gerganov <ggerganov@gmail.com> --------- Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
Configuration menu - View commit details
-
Copy full SHA for b07791a - Browse repository at this point
Copy the full SHA b07791aView commit details -
Configuration menu - View commit details
-
Copy full SHA for 7aeee88 - Browse repository at this point
Copy the full SHA 7aeee88View commit details
This comparison is taking too long to generate.
Unfortunately it looks like we can’t render this comparison for you right now. It might be too big, or there might be something weird with your repository.
You can try running this command locally to see the comparison on your machine:
git diff master@{1day}...master