@@ -1492,9 +1492,12 @@ int main(int argc, char ** argv) {
1492
1492
const cmd_params_instance * prev_inst = nullptr ;
1493
1493
1494
1494
int params_idx = 0 ;
1495
+ auto params_count = params_instances.size ();
1495
1496
for (const auto & inst : params_instances) {
1496
1497
params_idx ++;
1497
- LOG_TEE (" llama-bench: benchmark %d/%ld: starting\n " , params_idx, params_instances.size ());
1498
+ if (params.verbose ) {
1499
+ LOG_TEE (" llama-bench: benchmark %d/%ld: starting\n " , params_idx, params_count);
1500
+ }
1498
1501
// keep the same model between tests when possible
1499
1502
if (!lmodel || !prev_inst || !inst.equal_mparams (*prev_inst)) {
1500
1503
if (lmodel) {
@@ -1544,12 +1547,16 @@ int main(int argc, char ** argv) {
1544
1547
1545
1548
// warmup run
1546
1549
if (t.n_prompt > 0 ) {
1547
- LOG_TEE (" llama-bench: benchmark %d/%ld: warmup prompt run\n " , params_idx, params_instances.size ());
1550
+ if (params.verbose ) {
1551
+ LOG_TEE (" llama-bench: benchmark %d/%ld: warmup prompt run\n " , params_idx, params_count);
1552
+ }
1548
1553
// test_prompt(ctx, std::min(t.n_batch, std::min(t.n_prompt, 32)), 0, t.n_batch, t.n_threads);
1549
1554
test_prompt (ctx, t.n_prompt , 0 , t.n_batch , t.n_threads );
1550
1555
}
1551
1556
if (t.n_gen > 0 ) {
1552
- LOG_TEE (" llama-bench: benchmark %d/%ld: warmup generation run\n " , params_idx, params_instances.size ());
1557
+ if (params.verbose ) {
1558
+ LOG_TEE (" llama-bench: benchmark %d/%ld: warmup generation run\n " , params_idx, params_count);
1559
+ }
1553
1560
test_gen (ctx, 1 , 0 , t.n_threads );
1554
1561
}
1555
1562
@@ -1559,11 +1566,15 @@ int main(int argc, char ** argv) {
1559
1566
uint64_t t_start = get_time_ns ();
1560
1567
1561
1568
if (t.n_prompt > 0 ) {
1562
- LOG_TEE (" llama-bench: benchmark %d/%ld: prompt run %d/%d\n " , params_idx, params_instances.size (), i + 1 , params.reps );
1569
+ if (params.verbose ) {
1570
+ LOG_TEE (" llama-bench: benchmark %d/%ld: prompt run %d/%d\n " , params_idx, params_count, i + 1 , params.reps );
1571
+ }
1563
1572
test_prompt (ctx, t.n_prompt , 0 , t.n_batch , t.n_threads );
1564
1573
}
1565
1574
if (t.n_gen > 0 ) {
1566
- LOG_TEE (" llama-bench: benchmark %d/%ld: generation run %d/%d\n " , params_idx, params_instances.size (), i + 1 , params.reps );
1575
+ if (params.verbose ) {
1576
+ LOG_TEE (" llama-bench: benchmark %d/%ld: generation run %d/%d\n " , params_idx, params_count, i + 1 , params.reps );
1577
+ }
1567
1578
test_gen (ctx, t.n_gen , t.n_prompt , t.n_threads );
1568
1579
}
1569
1580
0 commit comments