[fix] fix completion stream api output_tokens not in usage (#3588)

This commit is contained in:
李泳桦
2025-08-25 18:31:57 +08:00
committed by GitHub
parent 4e369c7fa7
commit b1f8f1aa07

View File

@@ -413,6 +413,7 @@ class OpenAIServingCompletion:
usage=UsageInfo(
prompt_tokens=len(prompt_batched_token_ids[idx]),
completion_tokens=output_tokens[idx],
total_tokens=len(prompt_batched_token_ids[idx]) + output_tokens[idx],
),
)
yield f"data: {usage_chunk.model_dump_json(exclude_unset=True)}\n\n"