Skip to content

Commit 0173967

Browse files
authored
Merge pull request #2 from jamorphy/wip-orpheus-tts
cleanup
2 parents e004a88 + 1a6fa98 commit 0173967

File tree

3 files changed

+2
-7
lines changed

3 files changed

+2
-7
lines changed

examples/tts/orpheus-tts.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -298,7 +298,7 @@ int main(int argc, char **argv) {
298298
params.model = params.vocoder.model;
299299
params.n_batch = 2;
300300

301-
params.embedding = true
301+
params.embedding = true;
302302
// disable warmup, SNAC doesn't care about BOS or EOS tokens;
303303
params.warmup = false;
304304

src/llama-context.cpp

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -312,9 +312,7 @@ llama_context::llama_context(
312312

313313
// reserve pp graph first so that buffers are only allocated once
314314
{
315-
LLAMA_LOG_DEBUG("here 3\n");
316315
llama_ubatch ubatch_pp = { true, n_tokens, n_tokens / n_seqs, n_seqs, &token, nullptr, nullptr, nullptr, nullptr, nullptr};
317-
318316
auto * gf = graph_init();
319317
graph_build(ctx_compute.get(), gf, ubatch_pp, LLM_GRAPH_TYPE_DEFAULT);
320318
if (!ggml_backend_sched_reserve(sched.get(), gf)) {

src/llama-model.cpp

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1495,8 +1495,7 @@ bool llama_model::load_tensors(llama_model_loader & ml) {
14951495
if (flags & TENSOR_NOT_REQUIRED) {
14961496
return nullptr;
14971497
}
1498-
return nullptr;
1499-
//throw std::runtime_error(format("missing tensor '%s'", tn.str().c_str()));
1498+
throw std::runtime_error(format("missing tensor '%s'", tn.str().c_str()));
15001499
}
15011500

15021501
// some models use the token embedding tensor as the output, but since these are used in different layers and with different ops
@@ -1591,8 +1590,6 @@ bool llama_model::load_tensors(llama_model_loader & ml) {
15911590
return t;
15921591
}
15931592
}
1594-
fprintf(stderr, "create_tensor: Creating '%s' with ne=[%ld, %ld, %ld]\n",
1595-
tn_str.c_str(), ne.begin()[0], ne.begin()[1], ne.begin()[2]);
15961593
return ml.create_tensor(ctx, tn, ne, flags);
15971594
};
15981595

0 commit comments

Comments
 (0)