Skip to content

Commit 251ae54

Browse files
committed
feat: Auto-fill hparams.recurrent_layer_arr based on whether the model is recurrent
Branch: GraniteFour Signed-off-by: Gabe Goodhart <ghart@us.ibm.com>
1 parent bc8b9df commit 251ae54

File tree

1 file changed

+4
-0
lines changed

1 file changed

+4
-0
lines changed

src/llama-model.cpp

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -467,6 +467,10 @@ void llama_model::load_hparams(llama_model_loader & ml) {
467467
std::fill(hparams.n_head_arr.begin(), hparams.n_head_arr.end(), 0);
468468
std::fill(hparams.n_head_kv_arr.begin(), hparams.n_head_kv_arr.end(), 0);
469469
std::fill(hparams.n_ff_arr.begin(), hparams.n_ff_arr.end(), 0);
470+
std::fill(
471+
hparams.recurrent_layer_arr.begin(),
472+
hparams.recurrent_layer_arr.end(),
473+
llm_arch_is_recurrent(ml.get_arch()));
470474

471475
ml.get_key_or_arr(LLM_KV_FEED_FORWARD_LENGTH, hparams.n_ff_arr, hparams.n_layer, false);
472476
ml.get_key_or_arr(LLM_KV_ATTENTION_HEAD_COUNT, hparams.n_head_arr, hparams.n_layer, false);

0 commit comments

Comments
 (0)