mirror of
https://github.com/open-webui/open-webui
synced 2025-02-08 05:56:58 +00:00
refac
This commit is contained in:
parent
646671c576
commit
ba81b47800
@ -962,7 +962,7 @@ async def get_ollama_url(request: Request, model: str, url_idx: Optional[int] =
|
|||||||
)
|
)
|
||||||
url_idx = random.choice(models[model].get("urls", []))
|
url_idx = random.choice(models[model].get("urls", []))
|
||||||
url = request.app.state.config.OLLAMA_BASE_URLS[url_idx]
|
url = request.app.state.config.OLLAMA_BASE_URLS[url_idx]
|
||||||
return url
|
return url, url_idx
|
||||||
|
|
||||||
|
|
||||||
@router.post("/api/chat")
|
@router.post("/api/chat")
|
||||||
@ -1030,17 +1030,7 @@ async def generate_chat_completion(
|
|||||||
if ":" not in payload["model"]:
|
if ":" not in payload["model"]:
|
||||||
payload["model"] = f"{payload['model']}:latest"
|
payload["model"] = f"{payload['model']}:latest"
|
||||||
|
|
||||||
await get_all_models(request)
|
url, url_idx = await get_ollama_url(request, payload["model"], url_idx)
|
||||||
model = request.app.state.OLLAMA_MODELS.get(model_id)
|
|
||||||
if model:
|
|
||||||
url_idx = model["urls"].pop()
|
|
||||||
else:
|
|
||||||
raise HTTPException(
|
|
||||||
status_code=404,
|
|
||||||
detail="Model not found",
|
|
||||||
)
|
|
||||||
|
|
||||||
url = await get_ollama_url(request, payload["model"], url_idx)
|
|
||||||
api_config = request.app.state.config.OLLAMA_API_CONFIGS.get(
|
api_config = request.app.state.config.OLLAMA_API_CONFIGS.get(
|
||||||
str(url_idx),
|
str(url_idx),
|
||||||
request.app.state.config.OLLAMA_API_CONFIGS.get(url, {}), # Legacy support
|
request.app.state.config.OLLAMA_API_CONFIGS.get(url, {}), # Legacy support
|
||||||
@ -1142,7 +1132,7 @@ async def generate_openai_completion(
|
|||||||
if ":" not in payload["model"]:
|
if ":" not in payload["model"]:
|
||||||
payload["model"] = f"{payload['model']}:latest"
|
payload["model"] = f"{payload['model']}:latest"
|
||||||
|
|
||||||
url = await get_ollama_url(request, payload["model"], url_idx)
|
url, url_idx = await get_ollama_url(request, payload["model"], url_idx)
|
||||||
api_config = request.app.state.config.OLLAMA_API_CONFIGS.get(
|
api_config = request.app.state.config.OLLAMA_API_CONFIGS.get(
|
||||||
str(url_idx),
|
str(url_idx),
|
||||||
request.app.state.config.OLLAMA_API_CONFIGS.get(url, {}), # Legacy support
|
request.app.state.config.OLLAMA_API_CONFIGS.get(url, {}), # Legacy support
|
||||||
@ -1219,7 +1209,7 @@ async def generate_openai_chat_completion(
|
|||||||
if ":" not in payload["model"]:
|
if ":" not in payload["model"]:
|
||||||
payload["model"] = f"{payload['model']}:latest"
|
payload["model"] = f"{payload['model']}:latest"
|
||||||
|
|
||||||
url = await get_ollama_url(request, payload["model"], url_idx)
|
url, url_idx = await get_ollama_url(request, payload["model"], url_idx)
|
||||||
api_config = request.app.state.config.OLLAMA_API_CONFIGS.get(
|
api_config = request.app.state.config.OLLAMA_API_CONFIGS.get(
|
||||||
str(url_idx),
|
str(url_idx),
|
||||||
request.app.state.config.OLLAMA_API_CONFIGS.get(url, {}), # Legacy support
|
request.app.state.config.OLLAMA_API_CONFIGS.get(url, {}), # Legacy support
|
||||||
|
Loading…
Reference in New Issue
Block a user