Skip to content

Tags: BodhiSearch/llama.cpp

Tags

server-bodhiapp_20250817-b6184-fb7b493

Toggle server-bodhiapp_20250817-b6184-fb7b493's commit message
[Amir] changes for bodhiapp

- having git workflows to build binaries for supported platforms
- added ubuntu-arm64 platform to build

server-bodhiapp_20250817-b6184-6f884f1

Toggle server-bodhiapp_20250817-b6184-6f884f1's commit message
[Amir] changes for bodhiapp

- having git workflows to build binaries for supported platforms

server-bodhiapp_20250806-b6098-6186d1a

Toggle server-bodhiapp_20250806-b6098-6186d1a's commit message
[Amir] changes for bodhiapp

- having git workflows to build binaries for supported platforms

server-bodhiapp_20250727-b5999-95f47c6

Toggle server-bodhiapp_20250727-b5999-95f47c6's commit message
[Amir] changes for bodhiapp

- having git workflows to build binaries for supported platforms

server-bodhiapp_20250626-b5759-fe7ad85

Toggle server-bodhiapp_20250626-b5759-fe7ad85's commit message
[Amir] only building -cpu variants

server-bodhiapp_20250626-b5758-90f2a5f

Toggle server-bodhiapp_20250626-b5758-90f2a5f's commit message
[Amir] changes for bodhiapp

- having git workflows to build binaries for supported platforms

server-bodhiapp_20250626-b5758-40bbc4f

Toggle server-bodhiapp_20250626-b5758-40bbc4f's commit message
[Amir] changes for bodhiapp

- having git workflows to build binaries for supported platforms

server-bodhiapp_20250626-b5758-1cd8efd

Toggle server-bodhiapp_20250626-b5758-1cd8efd's commit message
[Amir] changes for bodhiapp

- having git workflows to build binaries for supported platforms

b5608

Toggle b5608's commit message

Verified

This commit was created on GitHub.com and signed with GitHub’s verified signature.
add geglu activation function (ggml-org#14074)

Co-authored-by: dinhhuy <huy.dinh@brains-tech.co.jp>

build-e997375d

Toggle build-e997375d's commit message
[BodhiApp] changes for bodhiapp.

[21-dec-24] using prompt if passed in chat completions, and not using messages
- added python integration tests for changes in server
- having add_special as request param to pass to upstream, allows pre-formatted chat messages to not be formatted again
- modified workflow to download and cache the llama2-7b model used for integration testing
[17-jan-25] updated to latest llama.cpp
server.cpp had changes where ctx_server.vocab was used instead of the earlier ctx_server.ctx
[26-jan-25] merged examples/server/utils.hpp, llama.cpp is starting to support tools in the chat template, resolved rebase conflict by merging our logic of preformatted messages with new feature
adding the github workflow to create the artifacts based on platform and gpu architecture, for macos-cpu+metal, ubuntu-cpu+cuda, windows-cpu+cuda-12.4+11.6
also added for other ubuntu+windows cpu variants and gpu archs like vulkan etc., but disabled for now
[14-march-2025] - added ai-docs folder with cpu strategy
- added github  workflow for windows based on the strategy document