@@ -249,27 +249,29 @@ static void test_llama_kv_cache_hybrid_constructor() {
249
249
250
250
std::unique_ptr<llama_kv_cache_unified> u_cache (
251
251
new llama_kv_cache_unified (
252
- /* model */ *model,
253
- /* filter */ unified_filter,
254
- /* type_k */ GGML_TYPE_F32,
255
- /* type_v */ GGML_TYPE_F16,
256
- /* v_trans */ false ,
257
- /* offload */ false ,
258
- /* kv_size */ 10 ,
259
- /* padding */ 10 ,
260
- /* n_swa */ 0 ,
261
- /* swa_type */ LLAMA_SWA_TYPE_NONE
252
+ /* model */ *model,
253
+ /* filter */ unified_filter,
254
+ /* type_k */ GGML_TYPE_F32,
255
+ /* type_v */ GGML_TYPE_F16,
256
+ /* v_trans */ false ,
257
+ /* offload */ false ,
258
+ /* kv_size */ 10 ,
259
+ /* n_seq_max */ 1 ,
260
+ /* padding */ 10 ,
261
+ /* n_swa */ 0 ,
262
+ /* swa_type */ LLAMA_SWA_TYPE_NONE
262
263
)
263
264
);
264
265
auto * u_cache_ptr = u_cache.get ();
265
266
std::unique_ptr<llama_kv_cache_recurrent> r_cache (
266
267
new llama_kv_cache_recurrent (
267
- /* model */ *model,
268
- /* filter */ recurrent_filter,
269
- /* type_k */ GGML_TYPE_F32,
270
- /* type_v */ GGML_TYPE_F16,
271
- /* offload */ false ,
272
- /* kv_size */ 10
268
+ /* model */ *model,
269
+ /* filter */ recurrent_filter,
270
+ /* type_k */ GGML_TYPE_F32,
271
+ /* type_v */ GGML_TYPE_F16,
272
+ /* offload */ false ,
273
+ /* kv_size */ 10 ,
274
+ /* n_seq_max */ 1
273
275
)
274
276
);
275
277
auto * r_cache_ptr = r_cache.get ();
0 commit comments