@@ -1491,10 +1491,10 @@ int main(int argc, char ** argv) {
14911491 llama_model * lmodel = nullptr ;
14921492 const cmd_params_instance * prev_inst = nullptr ;
14931493
1494- int params_idx = 1 ;
1494+ int params_idx = 0 ;
14951495 for (const auto & inst : params_instances) {
1496- LOG_TEE (" llama-bench: starting benchmark %d/%ld\n " , params_idx, params_instances.size ());
14971496 params_idx ++;
1497+ LOG_TEE (" llama-bench: benchmark %d/%ld: starting\n " , params_idx, params_instances.size ());
14981498 // keep the same model between tests when possible
14991499 if (!lmodel || !prev_inst || !inst.equal_mparams (*prev_inst)) {
15001500 if (lmodel) {
@@ -1544,10 +1544,12 @@ int main(int argc, char ** argv) {
15441544
15451545 // warmup run
15461546 if (t.n_prompt > 0 ) {
1547+ LOG_TEE (" llama-bench: benchmark %d/%ld: warmup prompt run\n " , params_idx, params_instances.size ());
15471548 // test_prompt(ctx, std::min(t.n_batch, std::min(t.n_prompt, 32)), 0, t.n_batch, t.n_threads);
15481549 test_prompt (ctx, t.n_prompt , 0 , t.n_batch , t.n_threads );
15491550 }
15501551 if (t.n_gen > 0 ) {
1552+ LOG_TEE (" llama-bench: benchmark %d/%ld: warmup generation run\n " , params_idx, params_instances.size ());
15511553 test_gen (ctx, 1 , 0 , t.n_threads );
15521554 }
15531555
@@ -1557,9 +1559,11 @@ int main(int argc, char ** argv) {
15571559 uint64_t t_start = get_time_ns ();
15581560
15591561 if (t.n_prompt > 0 ) {
1562+ LOG_TEE (" llama-bench: benchmark %d/%ld: prompt run %d/%d\n " , params_idx, params_instances.size (), i + 1 , params.reps );
15601563 test_prompt (ctx, t.n_prompt , 0 , t.n_batch , t.n_threads );
15611564 }
15621565 if (t.n_gen > 0 ) {
1566+ LOG_TEE (" llama-bench: benchmark %d/%ld: generation run %d/%d\n " , params_idx, params_instances.size (), i + 1 , params.reps );
15631567 test_gen (ctx, t.n_gen , t.n_prompt , t.n_threads );
15641568 }
15651569
0 commit comments