Skip to content

Commit 7c96b13

Browse files
committed
Reorder types for llm_type
1 parent 56adda7 commit 7c96b13

File tree

8 files changed

+114
-2
lines changed

8 files changed

+114
-2
lines changed

.roomodes

Lines changed: 19 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,19 @@
1+
customModes:
2+
- slug: documentation-writer
3+
name: ✍️ Documentation Writer
4+
roleDefinition: |
5+
You are a technical documentation expert specializing in creating clear, comprehensive documentation for software projects. Your expertise includes:
6+
Writing clear, concise technical documentation
7+
Creating and maintaining README files, API documentation, and user guides
8+
Following documentation best practices and style guides
9+
Understanding code to accurately document its functionality
10+
Organizing documentation in a logical, easily navigable structure
11+
whenToUse: |
12+
Use this mode when you need to create, update, or improve technical documentation. Ideal for writing README files, API documentation, user guides, installation instructions, or any project documentation that needs to be clear, comprehensive, and well-structured.
13+
description: Create clear technical project documentation
14+
groups:
15+
- read
16+
- edit
17+
- command
18+
customInstructions: |
19+
Focus on creating documentation that is clear, concise, and follows a consistent style. Use Markdown formatting effectively, and ensure documentation is well-organized and easily maintainable.

bld.sh

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,2 @@
1+
#!/bin/bash
2+
cmake --build build --config Release -j 8

embeddings

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,2 @@
1+
#!/bin/bash
2+
llama-server -m /mnt/win/k/models/Casual-Autopsy/snowflake-arctic-embed-l-v2.0-gguf/snowflake-arctic-embed-l-v2.0-q8_0.gguf -fa -a text-embedding-snowflake-arctic-embed-l-v2.0 --port 2234

polaris.jinja

Lines changed: 85 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,85 @@
1+
{%- if tools %}
2+
{{- '<|im_start|>system\n' }}
3+
{%- if messages[0].role == 'system' %}
4+
{{- messages[0].content + '\n\n' }}
5+
{%- endif %}
6+
{{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
7+
{%- for tool in tools %}
8+
{{- "\n" }}
9+
{{- tool | tojson }}
10+
{%- endfor %}
11+
{{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
12+
{%- else %}
13+
{%- if messages[0].role == 'system' %}
14+
{{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15+
{%- endif %}
16+
{%- endif %}
17+
{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
18+
{%- for message in messages[::-1] %}
19+
{%- set index = (messages|length - 1) - loop.index0 %}
20+
{%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}
21+
{%- set ns.multi_step_tool = false %}
22+
{%- set ns.last_query_index = index %}
23+
{%- endif %}
24+
{%- endfor %}
25+
{%- for message in messages %}
26+
{%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
27+
{{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
28+
{%- elif message.role == "assistant" %}
29+
{%- set content = message.content %}
30+
{%- set reasoning_content = '' %}
31+
{%- if message.reasoning_content is defined and message.reasoning_content is not none %}
32+
{%- set reasoning_content = message.reasoning_content %}
33+
{%- else %}
34+
{%- if '</think>' in message.content %}
35+
{%- set content = message.content.split('</think>')[-1].lstrip('\n') %}
36+
{%- set reasoning_content = message.content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
37+
{%- endif %}
38+
{%- endif %}
39+
{%- if loop.index0 > ns.last_query_index %}
40+
{%- if loop.last or (not loop.last and reasoning_content) %}
41+
{{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content.strip('\n') + '\n</think>\n\n' + content.lstrip('\n') }}
42+
{%- else %}
43+
{{- '<|im_start|>' + message.role + '\n' + content }}
44+
{%- endif %}
45+
{%- else %}
46+
{{- '<|im_start|>' + message.role + '\n' + content }}
47+
{%- endif %}
48+
{%- if message.tool_calls %}
49+
{%- for tool_call in message.tool_calls %}
50+
{%- if (loop.first and content) or (not loop.first) %}
51+
{{- '\n' }}
52+
{%- endif %}
53+
{%- if tool_call.function %}
54+
{%- set tool_call = tool_call.function %}
55+
{%- endif %}
56+
{{- '<tool_call>\n{"name": "' }}
57+
{{- tool_call.name }}
58+
{{- '", "arguments": ' }}
59+
{%- if tool_call.arguments is string %}
60+
{{- tool_call.arguments }}
61+
{%- else %}
62+
{{- tool_call.arguments | tojson }}
63+
{%- endif %}
64+
{{- '}\n</tool_call>' }}
65+
{%- endfor %}
66+
{%- endif %}
67+
{{- '<|im_end|>\n' }}
68+
{%- elif message.role == "tool" %}
69+
{%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
70+
{{- '<|im_start|>user' }}
71+
{%- endif %}
72+
{{- '\n<tool_response>\n' }}
73+
{{- message.content }}
74+
{{- '\n</tool_response>' }}
75+
{%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
76+
{{- '<|im_end|>\n' }}
77+
{%- endif %}
78+
{%- endif %}
79+
{%- endfor %}
80+
{%- if add_generation_prompt %}
81+
{{- '<|im_start|>assistant\n' }}
82+
{%- if enable_thinking is defined and enable_thinking is false %}
83+
{{- '<think>\n\n</think>\n\n' }}
84+
{%- endif %}
85+
{%- endif %}

prebld.sh

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
cmake -B build -DGGML_CUDA=ON -DGGML_CUDA_FORCE_CUBLAS=OFF -DGGML_CUDA_FORCE_MMQ=OFF -DGGML_CUDA_FA_ALL_QUANTS=1 -DCMAKE_CUDA_ARCHITECTURES=86 -DGGML_BLAS=ON -DGGML_BLAS_VENDOR=OpenBLAS -DCMAKE_BUILD_TYPE=Release

src/llama-model.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18071,7 +18071,7 @@ struct llm_build_seed_oss : public llm_graph_context {
1807118071
cur = build_norm(ffn_inp,
1807218072
model.layers[il].attn_post_norm, NULL,
1807318073
LLM_NORM_RMS, il);
18074-
cb(cur, "ffn_norm", il);
18074+
cb(cur, "attn_post_norm", il);
1807518075

1807618076
cur = build_ffn(cur,
1807718077
model.layers[il].ffn_up, NULL, NULL,

src/llama-model.h

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -76,6 +76,7 @@ enum llm_type {
7676
LLM_TYPE_32B,
7777
LLM_TYPE_34B,
7878
LLM_TYPE_35B,
79+
LLM_TYPE_36B, // Seed OSS
7980
LLM_TYPE_40B,
8081
LLM_TYPE_65B,
8182
LLM_TYPE_70B,
@@ -109,7 +110,6 @@ enum llm_type {
109110
LLM_TYPE_355B_A32B, // GLM-4.5
110111
LLM_TYPE_E2B,
111112
LLM_TYPE_E4B,
112-
LLM_TYPE_36B // Seed OSS
113113
};
114114

115115
std::string llama_rope_scaling_type_name(llama_rope_scaling_type rope_scaling_type);

up.sh

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,3 @@
1+
#!/bin/bash
2+
cd /devel/tools/llama.cpp
3+
git fetch && git pull && ./prebld.sh && ./bld.sh

0 commit comments

Comments
 (0)