Skip to content

webui : display prompt timing stats #16902

@ggerganov

Description

@ggerganov

Great. Btw, we could also show the stats for the prompt, not just for the generated tokens. This ok for now - we can expand this information in the future.

Originally posted by @ggerganov in #16901 (review)

Display the prompt_ stats in a similar way as the predicted_ stats:

json base = {
{"cache_n", cache_n},
{"prompt_n", prompt_n},
{"prompt_ms", prompt_ms},
{"prompt_per_token_ms", prompt_per_token_ms},
{"prompt_per_second", prompt_per_second},
{"predicted_n", predicted_n},
{"predicted_ms", predicted_ms},
{"predicted_per_token_ms", predicted_per_token_ms},
{"predicted_per_second", predicted_per_second},
};

Metadata

Metadata

Assignees

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions