fix: pipelines

This commit is contained in:
Timothy J. Baek 2024-05-31 13:30:12 -07:00
parent fc2b314c4f
commit cb8c45d864
4 changed files with 11 additions and 8 deletions

View File

@ -400,11 +400,6 @@ async def proxy(path: str, request: Request, user=Depends(get_verified_user)):
if "pipeline" in model and model.get("pipeline"): if "pipeline" in model and model.get("pipeline"):
payload["user"] = {"name": user.name, "id": user.id} payload["user"] = {"name": user.name, "id": user.id}
payload["title"] = (
True
if payload["stream"] == False and payload["max_tokens"] == 50
else False
)
# Check if the model is "gpt-4-vision-preview" and set "max_tokens" to 4000 # Check if the model is "gpt-4-vision-preview" and set "max_tokens" to 4000
# This is a workaround until OpenAI fixes the issue with this model # This is a workaround until OpenAI fixes the issue with this model

View File

@ -315,9 +315,13 @@ class PipelineMiddleware(BaseHTTPMiddleware):
else: else:
pass pass
if "pipeline" not in app.state.MODELS[model_id]:
if "chat_id" in data: if "chat_id" in data:
del data["chat_id"] del data["chat_id"]
if "title" in data:
del data["title"]
modified_body_bytes = json.dumps(data).encode("utf-8") modified_body_bytes = json.dumps(data).encode("utf-8")
# Replace the request body with the modified one # Replace the request body with the modified one
request._body = modified_body_bytes request._body = modified_body_bytes

View File

@ -336,6 +336,7 @@ export const generateTitle = async (
template: string, template: string,
model: string, model: string,
prompt: string, prompt: string,
chat_id?: string,
url: string = OPENAI_API_BASE_URL url: string = OPENAI_API_BASE_URL
) => { ) => {
let error = null; let error = null;
@ -361,7 +362,9 @@ export const generateTitle = async (
], ],
stream: false, stream: false,
// Restricting the max tokens to 50 to avoid long titles // Restricting the max tokens to 50 to avoid long titles
max_tokens: 50 max_tokens: 50,
...(chat_id && { chat_id: chat_id }),
title: true
}) })
}) })
.then(async (res) => { .then(async (res) => {

View File

@ -1118,6 +1118,7 @@
) + ' {{prompt}}', ) + ' {{prompt}}',
titleModelId, titleModelId,
userPrompt, userPrompt,
$chatId,
titleModel?.owned_by === 'openai' ?? false titleModel?.owned_by === 'openai' ?? false
? `${OPENAI_API_BASE_URL}` ? `${OPENAI_API_BASE_URL}`
: `${OLLAMA_API_BASE_URL}/v1` : `${OLLAMA_API_BASE_URL}/v1`