Skip to content

Bench token generation at long context sizes #10936

Answered by ggerganov
Mushoz asked this question in Q&A
Discussion options

You must be logged in to vote

You can get these numbers with the llama-batched-bench tool, although it does not compute uncertainties like the llama-bench tool.

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@Mushoz
Comment options

Answer selected by Mushoz
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants