mirror of
https://github.com/open-webui/open-webui
synced 2025-05-18 20:31:36 +00:00
Merge pull request #10360 from ferret99gt/ollama-option-conversion-fixes
fix: Update Ollama option handling in payload.py's convert_payload_openai_to_ollama
This commit is contained in:
commit
13e58c1b64
@ -212,34 +212,19 @@ def convert_payload_openai_to_ollama(openai_payload: dict) -> dict:
|
|||||||
ollama_payload["format"] = openai_payload["format"]
|
ollama_payload["format"] = openai_payload["format"]
|
||||||
|
|
||||||
# If there are advanced parameters in the payload, format them in Ollama's options field
|
# If there are advanced parameters in the payload, format them in Ollama's options field
|
||||||
ollama_options = {}
|
|
||||||
|
|
||||||
if openai_payload.get("options"):
|
if openai_payload.get("options"):
|
||||||
ollama_payload["options"] = openai_payload["options"]
|
ollama_payload["options"] = openai_payload["options"]
|
||||||
ollama_options = openai_payload["options"]
|
ollama_options = openai_payload["options"]
|
||||||
|
|
||||||
# Handle parameters which map directly
|
# Re-Mapping OpenAI's `max_tokens` -> Ollama's `num_predict`
|
||||||
for param in ["temperature", "top_p", "seed"]:
|
if "max_tokens" in ollama_options:
|
||||||
if param in openai_payload:
|
ollama_options["num_predict"] = ollama_options["max_tokens"]
|
||||||
ollama_options[param] = openai_payload[param]
|
del ollama_options["max_tokens"] # To prevent Ollama warning of invalid option provided
|
||||||
|
|
||||||
# Mapping OpenAI's `max_tokens` -> Ollama's `num_predict`
|
# Ollama lacks a "system" prompt option. It has to be provided as a direct parameter, so we copy it down.
|
||||||
if "max_completion_tokens" in openai_payload:
|
if "system" in ollama_options:
|
||||||
ollama_options["num_predict"] = openai_payload["max_completion_tokens"]
|
ollama_payload["system"] = ollama_options["system"]
|
||||||
elif "max_tokens" in openai_payload:
|
del ollama_options["system"] # To prevent Ollama warning of invalid option provided
|
||||||
ollama_options["num_predict"] = openai_payload["max_tokens"]
|
|
||||||
|
|
||||||
# Handle frequency / presence_penalty, which needs renaming and checking
|
|
||||||
if "frequency_penalty" in openai_payload:
|
|
||||||
ollama_options["repeat_penalty"] = openai_payload["frequency_penalty"]
|
|
||||||
|
|
||||||
if "presence_penalty" in openai_payload and "penalty" not in ollama_options:
|
|
||||||
# We are assuming presence penalty uses a similar concept in Ollama, which needs custom handling if exists.
|
|
||||||
ollama_options["new_topic_penalty"] = openai_payload["presence_penalty"]
|
|
||||||
|
|
||||||
# Add options to payload if any have been set
|
|
||||||
if ollama_options:
|
|
||||||
ollama_payload["options"] = ollama_options
|
|
||||||
|
|
||||||
if "metadata" in openai_payload:
|
if "metadata" in openai_payload:
|
||||||
ollama_payload["metadata"] = openai_payload["metadata"]
|
ollama_payload["metadata"] = openai_payload["metadata"]
|
||||||
|
Loading…
Reference in New Issue
Block a user