We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 56c4f7b commit 5280749Copy full SHA for 5280749
src/llama.cpp
@@ -5353,7 +5353,7 @@ static void llm_load_hparams(
5353
default: model.type = e_model::MODEL_UNKNOWN;
5354
}
5355
} break;
5356
- case LLM_ARCH_RWKV:
+ case LLM_ARCH_RWKV:
5357
{
5358
ml.get_key(LLM_KV_ATTENTION_LAYERNORM_EPS, hparams.f_norm_eps);
5359
ml.get_key(LLM_KV_WKV_HEAD_SIZE, hparams.wkv_head_size);
@@ -8745,7 +8745,6 @@ static struct ggml_tensor * llm_build_channel_mix(
8745
const struct llama_layer * layer,
8746
struct ggml_tensor * current,
8747
struct ggml_tensor * x_prev) {
8748
-
8749
struct ggml_tensor * sx = ggml_sub(ctx, x_prev, current);
8750
struct ggml_tensor * xk = ggml_add_inplace(
8751
ctx,
0 commit comments