-
Notifications
You must be signed in to change notification settings - Fork 13.6k
Open
Labels
Description
Great. Btw, we could also show the stats for the prompt, not just for the generated tokens. This ok for now - we can expand this information in the future.
Originally posted by @ggerganov in #16901 (review)
Display the prompt_ stats in a similar way as the predicted_ stats:
llama.cpp/tools/server/server.cpp
Lines 639 to 651 in 4621d99
| json base = { | |
| {"cache_n", cache_n}, | |
| {"prompt_n", prompt_n}, | |
| {"prompt_ms", prompt_ms}, | |
| {"prompt_per_token_ms", prompt_per_token_ms}, | |
| {"prompt_per_second", prompt_per_second}, | |
| {"predicted_n", predicted_n}, | |
| {"predicted_ms", predicted_ms}, | |
| {"predicted_per_token_ms", predicted_per_token_ms}, | |
| {"predicted_per_second", predicted_per_second}, | |
| }; |
ServeurpersoCom