Skip to content

Commit 325afb3

Browse files
authored
llama: fix missing k_cache store for rwkv6qwen2 (ggml-org#11445)
Signed-off-by: Molly Sophia <[email protected]>
1 parent 794fe23 commit 325afb3

File tree

1 file changed

+3
-7
lines changed

1 file changed

+3
-7
lines changed

src/llama.cpp

+3-7
Original file line numberDiff line numberDiff line change
@@ -7700,17 +7700,13 @@ struct llm_build_context {
77007700
1
77017701
);
77027702

7703+
struct ggml_tensor * last_norm_att = ggml_view_3d(ctx0, x_norm_att, n_embd, 1, n_seqs, x_norm_att->nb[1], x_norm_att->nb[2], (n_seq_tokens-1)*n_embd*ggml_element_size(x_norm_att));
77037704
ggml_build_forward_expand(
77047705
gf,
77057706
ggml_cpy(
77067707
ctx0,
7707-
wkv_states,
7708-
ggml_view_1d(
7709-
ctx0,
7710-
kv_self.v_l[il],
7711-
hparams.n_embd_v_s() * n_seqs,
7712-
hparams.n_embd_v_s() * kv_head * ggml_element_size(kv_self.v_l[il])
7713-
)
7708+
ggml_view_1d(ctx0, last_norm_att, n_embd * n_seqs, 0),
7709+
ggml_view_1d(ctx0, kv_self.k_l[il], hparams.n_embd_k_s() * n_seqs, hparams.n_embd_k_s() * kv_head * ggml_element_size(kv_self.k_l[il]))
77147710
)
77157711
);
77167712

0 commit comments

Comments
 (0)