This repository was archived by the owner on Sep 12, 2024. It is now read-only.

Description
Hi,
In llama.cpp examples, the model prints loading times, tokens per sec and other useful info used for debug:
llama_print_timings: load time = 576.45 ms
llama_print_timings: sample time = 283.10 ms / 400 runs ( 0.71 ms per token, 1412.91 tokens per second)
llama_print_timings: prompt eval time = 599.83 ms / 19 tokens ( 31.57 ms per token, 31.68 tokens per second)
llama_print_timings: eval time = 24513.59 ms / 399 runs ( 61.44 ms per token, 16.28 tokens per second)
llama_print_timings: total time = 25431.49 ms
However llama-node doesn't print these. Is it a flag which should be turned on or what?