Skip to content

Commit 07fb71a

Browse files
committed
kv-cache : some comments
ggml-ci
1 parent 1875575 commit 07fb71a

File tree

2 files changed

+2
-1
lines changed

2 files changed

+2
-1
lines changed

src/llama-context.h

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -185,7 +185,7 @@ struct llama_context {
185185
ggml_cgraph * gf,
186186
bool batched);
187187

188-
// reserve a graph
188+
// reserve a graph with a dummy ubatch of the specified size
189189
ggml_cgraph * graph_reserve(uint32_t n_tokens, uint32_t n_seqs, uint32_t n_outputs);
190190

191191
private:

src/llama-kv-cache.h

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -135,6 +135,7 @@ class llama_kv_cache_unified : public llama_kv_cache {
135135
ggml_tensor * cpy_k(ggml_context * ctx, ggml_tensor * k_cur, int32_t il) const;
136136
ggml_tensor * cpy_v(ggml_context * ctx, ggml_tensor * v_cur, int32_t il) const;
137137

138+
// find places for the provided ubatches in the cache, returns the head locations
138139
// return empty vector on failure
139140
std::vector<uint32_t> prepare(const std::vector<llama_ubatch> & ubatches);
140141

0 commit comments

Comments
 (0)