We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 3f3769b commit a75cb30Copy full SHA for a75cb30
src/llama-context.cpp
@@ -1050,6 +1050,9 @@ int llama_context::decode(llama_batch & inp_batch) {
1050
// finalize the batch processing
1051
kv_guard.commit();
1052
1053
+ // set to total number of outputs in the batch, for use in llama_get_logits_ith
1054
+ n_outputs = n_outputs_all;
1055
+
1056
// set output mappings
1057
{
1058
bool sorted_output = true;
@@ -1103,9 +1106,6 @@ int llama_context::decode(llama_batch & inp_batch) {
1103
1106
}
1104
1107
1105
1108
- // set to total number of outputs in the batch, for use in llama_get_logits_ith
- n_outputs = n_outputs_all;
-
1109
// wait for the computation to finish (automatically done when obtaining the model output)
1110
//synchronize();
1111
0 commit comments