@@ -1373,9 +1373,9 @@ struct llm_build_context {
1373
1373
inp = ggml_graph_node (gf, i);
1374
1374
if (strcmp (inp->name , " result_norm" ) == 0 || strcmp (inp->name , " result_embd" ) == 0 ) {
1375
1375
break ;
1376
- } else {
1377
- inp = nullptr ;
1378
1376
}
1377
+
1378
+ inp = nullptr ;
1379
1379
}
1380
1380
GGML_ASSERT (inp != nullptr && " missing result_norm/result_embd tensor" );
1381
1381
@@ -1431,7 +1431,7 @@ struct llm_build_context {
1431
1431
return gf;
1432
1432
}
1433
1433
1434
- struct ggml_tensor * llm_build_pos_bucket (bool causal) {
1434
+ struct ggml_tensor * build_pos_bucket (bool causal) {
1435
1435
if (causal) {
1436
1436
lctx.inp_pos_bucket = ggml_new_tensor_2d (ctx0, GGML_TYPE_I32, n_kv, n_tokens);
1437
1437
} else {
@@ -1444,7 +1444,7 @@ struct llm_build_context {
1444
1444
return lctx.inp_pos_bucket ;
1445
1445
}
1446
1446
1447
- struct ggml_tensor * llm_build_pos_bias (struct ggml_tensor * pos_bucket, struct ggml_tensor * attn_rel_b) {
1447
+ struct ggml_tensor * build_pos_bias (struct ggml_tensor * pos_bucket, struct ggml_tensor * attn_rel_b) {
1448
1448
struct ggml_tensor * pos_bucket_1d = ggml_view_1d (ctx0, pos_bucket, pos_bucket->ne [0 ] * pos_bucket->ne [1 ], 0 );
1449
1449
cb (pos_bucket_1d, " pos_bucket_1d" , -1 );
1450
1450
@@ -1463,15 +1463,15 @@ struct llm_build_context {
1463
1463
return pos_bias;
1464
1464
}
1465
1465
1466
- struct ggml_tensor * llm_build_inp_embd_enc () {
1466
+ struct ggml_tensor * build_inp_embd_enc () {
1467
1467
const int64_t n_embd = hparams.n_embd ;
1468
1468
lctx.inp_embd_enc = ggml_new_tensor_2d (ctx0, GGML_TYPE_F32, n_embd, n_outputs_enc);
1469
1469
ggml_set_input (lctx.inp_embd_enc );
1470
1470
cb (lctx.inp_embd_enc , " embd_enc" , -1 );
1471
1471
return lctx.inp_embd_enc ;
1472
1472
}
1473
1473
1474
- struct ggml_tensor * llm_build_inp_KQ_mask_cross () {
1474
+ struct ggml_tensor * build_inp_KQ_mask_cross () {
1475
1475
lctx.inp_KQ_mask_cross = ggml_new_tensor_2d (ctx0, GGML_TYPE_F32, n_outputs_enc, GGML_PAD (n_tokens, GGML_KQ_MASK_PAD));
1476
1476
ggml_set_input (lctx.inp_KQ_mask_cross );
1477
1477
cb (lctx.inp_KQ_mask_cross , " KQ_mask_cross" , -1 );
@@ -6775,7 +6775,7 @@ struct llm_build_context {
6775
6775
inpL = llm_build_inp_embd (ctx0, lctx, hparams, ubatch, model.tok_embd , cb);
6776
6776
6777
6777
GGML_ASSERT (lctx.is_encoding );
6778
- struct ggml_tensor * pos_bucket_enc = llm_build_pos_bucket (false );
6778
+ struct ggml_tensor * pos_bucket_enc = build_pos_bucket (false );
6779
6779
6780
6780
// KQ_mask (mask for 1 head, it will be broadcasted to all heads)
6781
6781
struct ggml_tensor * KQ_mask_enc = build_inp_KQ_mask (false );
@@ -6810,7 +6810,7 @@ struct llm_build_context {
6810
6810
cb (kq, " kq" , il);
6811
6811
6812
6812
struct ggml_tensor * attn_rel_b = model.layers [il].attn_rel_b_enc ? model.layers [il].attn_rel_b_enc : model.layers [0 ].attn_rel_b_enc ;
6813
- struct ggml_tensor * pos_bias = llm_build_pos_bias (pos_bucket_enc, attn_rel_b);
6813
+ struct ggml_tensor * pos_bias = build_pos_bias (pos_bucket_enc, attn_rel_b);
6814
6814
struct ggml_tensor * kq_b = ggml_add (ctx0, kq, pos_bias);
6815
6815
cb (kq_b, " kq_b" , il);
6816
6816
@@ -6909,11 +6909,11 @@ struct llm_build_context {
6909
6909
GGML_ASSERT (!lctx.is_encoding );
6910
6910
GGML_ASSERT (n_outputs_enc > 0 && " call llama_encode() first" );
6911
6911
6912
- struct ggml_tensor * embd_enc = llm_build_inp_embd_enc ();
6913
- struct ggml_tensor * pos_bucket_dec = llm_build_pos_bucket (true );
6912
+ struct ggml_tensor * embd_enc = build_inp_embd_enc ();
6913
+ struct ggml_tensor * pos_bucket_dec = build_pos_bucket (true );
6914
6914
6915
6915
struct ggml_tensor * KQ_mask_dec = build_inp_KQ_mask ();
6916
- struct ggml_tensor * KQ_mask_cross = llm_build_inp_KQ_mask_cross ();
6916
+ struct ggml_tensor * KQ_mask_cross = build_inp_KQ_mask_cross ();
6917
6917
6918
6918
for (int il = 0 ; il < n_layer; ++il) {
6919
6919
struct ggml_tensor * inpSA = inpL;
@@ -6961,7 +6961,7 @@ struct llm_build_context {
6961
6961
cb (kq, " kq" , il);
6962
6962
6963
6963
struct ggml_tensor * attn_rel_b = model.layers [il].attn_rel_b ? model.layers [il].attn_rel_b : model.layers [0 ].attn_rel_b ;
6964
- struct ggml_tensor * pos_bias = llm_build_pos_bias (pos_bucket_dec, attn_rel_b);
6964
+ struct ggml_tensor * pos_bias = build_pos_bias (pos_bucket_dec, attn_rel_b);
6965
6965
struct ggml_tensor * kq_b = ggml_add (ctx0, kq, pos_bias);
6966
6966
cb (kq_b, " kq_b" , il);
6967
6967
0 commit comments