Skip to content

Commit 43f3ce6

Browse files
committed
fix trailing whitespace
1 parent 85cf8bb commit 43f3ce6

File tree

1 file changed

+7
-8
lines changed

1 file changed

+7
-8
lines changed

src/llama.cpp

Lines changed: 7 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -7757,9 +7757,9 @@ static bool llm_load_tensors(
77577757
}
77587758
} break;
77597759
case LLM_ARCH_DECI:
7760-
{
7760+
{
77617761
model.tok_embd = create_tensor(tn(LLM_TENSOR_TOKEN_EMBD, "weight"), {n_embd, n_vocab}, 0);
7762-
7762+
77637763
// output
77647764
model.output_norm = create_tensor(tn(LLM_TENSOR_OUTPUT_NORM, "weight"), {n_embd}, 0);
77657765
model.output = create_tensor(tn(LLM_TENSOR_OUTPUT, "weight"), {n_embd, n_vocab}, llama_model_loader::TENSOR_NOT_REQUIRED);
@@ -11028,9 +11028,9 @@ struct llm_build_context {
1102811028
}
1102911029

1103011030
struct ggml_cgraph * build_deci() {
11031-
struct ggml_cgraph * gf = ggml_new_graph_custom(ctx0, llama_model_max_nodes(model), false);
11031+
struct ggml_cgraph * gf = ggml_new_graph_custom(ctx0, llama_model_max_nodes(model), false);
1103211032

11033-
// mutable variable, needed during the last layer of the computation to skip unused tokens
11033+
// mutable variable, needed during the last layer of the computation to skip unused tokens
1103411034
int32_t n_tokens = this->n_tokens;
1103511035

1103611036
const int64_t n_embd_head = hparams.n_embd_head_v;
@@ -11054,9 +11054,9 @@ struct llm_build_context {
1105411054
const int64_t n_head_kv = hparams.n_head_kv(il);
1105511055
const int64_t n_head = hparams.n_head(il);
1105611056

11057-
if (n_head == 0) // attention-free layer of Llama-3_1-Nemotron-51B
11057+
if (n_head == 0) { // attention-free layer of Llama-3_1-Nemotron-51B
1105811058
cur = inpL;
11059-
else {
11059+
} else {
1106011060
// norm
1106111061
cur = llm_build_norm(ctx0, inpL, hparams,
1106211062
model.layers[il].attn_norm, NULL,
@@ -11067,9 +11067,8 @@ struct llm_build_context {
1106711067
if (n_head > 0 && n_head_kv == 0) { // "linear attention" of Llama-3_1-Nemotron-51B
1106811068
cur = llm_build_lora_mm(lctx, ctx0, model.layers[il].wo, cur);
1106911069
cb(cur, "wo", il);
11070-
} else if (n_head > 0)
11070+
} else if (n_head > 0) {
1107111071
// self-attention
11072-
{
1107311072
// rope freq factors for llama3; may return nullptr for llama2 and other models
1107411073
struct ggml_tensor * rope_factors = build_rope_factors(il);
1107511074

0 commit comments

Comments
 (0)