Return Correct Prompt Eval Count Regardless of Cache Prompt (#5371)
* openai compatibility
* Revert "openai compatibility"
This reverts commit d3f98a811e.
* remove erroneous subtraction of prompt cache
This commit is contained in:
parent
daed0634a9
commit
3b5a4a77f3
|
|
@ -1732,7 +1732,7 @@ struct llama_server_context
|
|||
slot.n_past -= 1;
|
||||
}
|
||||
|
||||
slot.n_prompt_tokens_processed = slot.n_prompt_tokens - slot.n_past;
|
||||
slot.n_prompt_tokens_processed = slot.n_prompt_tokens;
|
||||
|
||||
if (slot.ga_n != 1)
|
||||
{
|
||||
|
|
|
|||
Loading…
Reference in New Issue