mirror of
https://github.com/open-webui/open-webui
synced 2024-11-16 13:40:55 +00:00
fix: pipelines
This commit is contained in:
parent
fc2b314c4f
commit
cb8c45d864
@ -400,11 +400,6 @@ async def proxy(path: str, request: Request, user=Depends(get_verified_user)):
|
|||||||
|
|
||||||
if "pipeline" in model and model.get("pipeline"):
|
if "pipeline" in model and model.get("pipeline"):
|
||||||
payload["user"] = {"name": user.name, "id": user.id}
|
payload["user"] = {"name": user.name, "id": user.id}
|
||||||
payload["title"] = (
|
|
||||||
True
|
|
||||||
if payload["stream"] == False and payload["max_tokens"] == 50
|
|
||||||
else False
|
|
||||||
)
|
|
||||||
|
|
||||||
# Check if the model is "gpt-4-vision-preview" and set "max_tokens" to 4000
|
# Check if the model is "gpt-4-vision-preview" and set "max_tokens" to 4000
|
||||||
# This is a workaround until OpenAI fixes the issue with this model
|
# This is a workaround until OpenAI fixes the issue with this model
|
||||||
|
@ -315,9 +315,13 @@ class PipelineMiddleware(BaseHTTPMiddleware):
|
|||||||
else:
|
else:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
if "pipeline" not in app.state.MODELS[model_id]:
|
||||||
if "chat_id" in data:
|
if "chat_id" in data:
|
||||||
del data["chat_id"]
|
del data["chat_id"]
|
||||||
|
|
||||||
|
if "title" in data:
|
||||||
|
del data["title"]
|
||||||
|
|
||||||
modified_body_bytes = json.dumps(data).encode("utf-8")
|
modified_body_bytes = json.dumps(data).encode("utf-8")
|
||||||
# Replace the request body with the modified one
|
# Replace the request body with the modified one
|
||||||
request._body = modified_body_bytes
|
request._body = modified_body_bytes
|
||||||
|
@ -336,6 +336,7 @@ export const generateTitle = async (
|
|||||||
template: string,
|
template: string,
|
||||||
model: string,
|
model: string,
|
||||||
prompt: string,
|
prompt: string,
|
||||||
|
chat_id?: string,
|
||||||
url: string = OPENAI_API_BASE_URL
|
url: string = OPENAI_API_BASE_URL
|
||||||
) => {
|
) => {
|
||||||
let error = null;
|
let error = null;
|
||||||
@ -361,7 +362,9 @@ export const generateTitle = async (
|
|||||||
],
|
],
|
||||||
stream: false,
|
stream: false,
|
||||||
// Restricting the max tokens to 50 to avoid long titles
|
// Restricting the max tokens to 50 to avoid long titles
|
||||||
max_tokens: 50
|
max_tokens: 50,
|
||||||
|
...(chat_id && { chat_id: chat_id }),
|
||||||
|
title: true
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
.then(async (res) => {
|
.then(async (res) => {
|
||||||
|
@ -1118,6 +1118,7 @@
|
|||||||
) + ' {{prompt}}',
|
) + ' {{prompt}}',
|
||||||
titleModelId,
|
titleModelId,
|
||||||
userPrompt,
|
userPrompt,
|
||||||
|
$chatId,
|
||||||
titleModel?.owned_by === 'openai' ?? false
|
titleModel?.owned_by === 'openai' ?? false
|
||||||
? `${OPENAI_API_BASE_URL}`
|
? `${OPENAI_API_BASE_URL}`
|
||||||
: `${OLLAMA_API_BASE_URL}/v1`
|
: `${OLLAMA_API_BASE_URL}/v1`
|
||||||
|
Loading…
Reference in New Issue
Block a user