diff --git a/backend/open_webui/utils/response.py b/backend/open_webui/utils/response.py index 891016e43..d429db8aa 100644 --- a/backend/open_webui/utils/response.py +++ b/backend/open_webui/utils/response.py @@ -29,7 +29,7 @@ async def convert_streaming_response_ollama_to_openai(ollama_streaming_response) ( ( data.get("eval_count", 0) - / ((data.get("eval_duration", 0) / 1_000_000_000)) + / ((data.get("eval_duration", 0) / 1_000_000)) ) * 100 ), @@ -43,12 +43,7 @@ async def convert_streaming_response_ollama_to_openai(ollama_streaming_response) ( ( data.get("prompt_eval_count", 0) - / ( - ( - data.get("prompt_eval_duration", 0) - / 1_000_000_000 - ) - ) + / ((data.get("prompt_eval_duration", 0) / 1_000_000)) ) * 100 ), @@ -57,20 +52,12 @@ async def convert_streaming_response_ollama_to_openai(ollama_streaming_response) if data.get("prompt_eval_duration", 0) > 0 else "N/A" ), - "total_duration": round( - ((data.get("total_duration", 0) / 1_000_000) * 100), 2 - ), - "load_duration": round( - ((data.get("load_duration", 0) / 1_000_000) * 100), 2 - ), + "total_duration": data.get("total_duration", 0), + "load_duration": data.get("load_duration", 0), "prompt_eval_count": data.get("prompt_eval_count", 0), - "prompt_eval_duration": round( - ((data.get("prompt_eval_duration", 0) / 1_000_000) * 100), 2 - ), + "prompt_eval_duration": data.get("prompt_eval_duration", 0), "eval_count": data.get("eval_count", 0), - "eval_duration": round( - ((data.get("eval_duration", 0) / 1_000_000) * 100), 2 - ), + "eval_duration": data.get("eval_duration", 0), "approximate_total": ( lambda s: f"{s // 3600}h{(s % 3600) // 60}m{s % 60}s" )((data.get("total_duration", 0) or 0) // 1_000_000_000),