mirror of
https://github.com/open-webui/open-webui
synced 2024-11-08 17:49:57 +00:00
722 lines
19 KiB
Svelte
722 lines
19 KiB
Svelte
<script lang="ts">
|
|
import { v4 as uuidv4 } from 'uuid';
|
|
import toast from 'svelte-french-toast';
|
|
|
|
import { onMount, tick } from 'svelte';
|
|
import { goto } from '$app/navigation';
|
|
import { page } from '$app/stores';
|
|
|
|
import { models, modelfiles, user, settings, chats, chatId, config } from '$lib/stores';
|
|
import { copyToClipboard, splitStream } from '$lib/utils';
|
|
|
|
import { generateChatCompletion, generateTitle } from '$lib/apis/ollama';
|
|
import { createNewChat, getChatList, updateChatById } from '$lib/apis/chats';
|
|
import { queryVectorDB } from '$lib/apis/rag';
|
|
import { generateOpenAIChatCompletion } from '$lib/apis/openai';
|
|
|
|
import MessageInput from '$lib/components/chat/MessageInput.svelte';
|
|
import Messages from '$lib/components/chat/Messages.svelte';
|
|
import ModelSelector from '$lib/components/chat/ModelSelector.svelte';
|
|
import Navbar from '$lib/components/layout/Navbar.svelte';
|
|
import { RAGTemplate } from '$lib/utils/rag';
|
|
|
|
let stopResponseFlag = false;
|
|
let autoScroll = true;
|
|
let processing = '';
|
|
|
|
let selectedModels = [''];
|
|
|
|
let selectedModelfile = null;
|
|
$: selectedModelfile =
|
|
selectedModels.length === 1 &&
|
|
$modelfiles.filter((modelfile) => modelfile.tagName === selectedModels[0]).length > 0
|
|
? $modelfiles.filter((modelfile) => modelfile.tagName === selectedModels[0])[0]
|
|
: null;
|
|
|
|
let selectedModelfiles = {};
|
|
$: selectedModelfiles = selectedModels.reduce((a, tagName, i, arr) => {
|
|
const modelfile =
|
|
$modelfiles.filter((modelfile) => modelfile.tagName === tagName)?.at(0) ?? undefined;
|
|
|
|
return {
|
|
...a,
|
|
...(modelfile && { [tagName]: modelfile })
|
|
};
|
|
}, {});
|
|
|
|
let chat = null;
|
|
|
|
let title = '';
|
|
let prompt = '';
|
|
let files = [];
|
|
let messages = [];
|
|
let history = {
|
|
messages: {},
|
|
currentId: null
|
|
};
|
|
|
|
$: if (history.currentId !== null) {
|
|
let _messages = [];
|
|
|
|
let currentMessage = history.messages[history.currentId];
|
|
while (currentMessage !== null) {
|
|
_messages.unshift({ ...currentMessage });
|
|
currentMessage =
|
|
currentMessage.parentId !== null ? history.messages[currentMessage.parentId] : null;
|
|
}
|
|
messages = _messages;
|
|
} else {
|
|
messages = [];
|
|
}
|
|
|
|
onMount(async () => {
|
|
await initNewChat();
|
|
});
|
|
|
|
//////////////////////////
|
|
// Web functions
|
|
//////////////////////////
|
|
|
|
const initNewChat = async () => {
|
|
window.history.replaceState(history.state, '', `/`);
|
|
|
|
console.log('initNewChat');
|
|
|
|
await chatId.set('');
|
|
console.log($chatId);
|
|
|
|
autoScroll = true;
|
|
|
|
title = '';
|
|
messages = [];
|
|
history = {
|
|
messages: {},
|
|
currentId: null
|
|
};
|
|
|
|
console.log($config);
|
|
|
|
if ($page.url.searchParams.get('models')) {
|
|
selectedModels = $page.url.searchParams.get('models')?.split(',');
|
|
} else if ($settings?.models) {
|
|
selectedModels = $settings?.models;
|
|
} else if ($config?.default_models) {
|
|
selectedModels = $config?.default_models.split(',');
|
|
} else {
|
|
selectedModels = [''];
|
|
}
|
|
|
|
let _settings = JSON.parse(localStorage.getItem('settings') ?? '{}');
|
|
settings.set({
|
|
..._settings
|
|
});
|
|
};
|
|
|
|
//////////////////////////
|
|
// Ollama functions
|
|
//////////////////////////
|
|
|
|
const submitPrompt = async (userPrompt, _user = null) => {
|
|
console.log('submitPrompt', $chatId);
|
|
|
|
if (selectedModels.includes('')) {
|
|
toast.error('Model not selected');
|
|
} else if (messages.length != 0 && messages.at(-1).done != true) {
|
|
// Response not done
|
|
console.log('wait');
|
|
} else if (
|
|
files.length > 0 &&
|
|
files.filter((file) => file.upload_status === false).length > 0
|
|
) {
|
|
// Upload not done
|
|
toast.error(
|
|
`Oops! Hold tight! Your files are still in the processing oven. We're cooking them up to perfection. Please be patient and we'll let you know once they're ready.`
|
|
);
|
|
} else {
|
|
// Reset chat message textarea height
|
|
document.getElementById('chat-textarea').style.height = '';
|
|
|
|
// Create user message
|
|
let userMessageId = uuidv4();
|
|
let userMessage = {
|
|
id: userMessageId,
|
|
parentId: messages.length !== 0 ? messages.at(-1).id : null,
|
|
childrenIds: [],
|
|
role: 'user',
|
|
user: _user ?? undefined,
|
|
content: userPrompt,
|
|
files: files.length > 0 ? files : undefined,
|
|
timestamp: Math.floor(Date.now() / 1000) // Unix epoch
|
|
};
|
|
|
|
// Add message to history and Set currentId to messageId
|
|
history.messages[userMessageId] = userMessage;
|
|
history.currentId = userMessageId;
|
|
|
|
// Append messageId to childrenIds of parent message
|
|
if (messages.length !== 0) {
|
|
history.messages[messages.at(-1).id].childrenIds.push(userMessageId);
|
|
}
|
|
|
|
// Wait until history/message have been updated
|
|
await tick();
|
|
|
|
// Create new chat if only one message in messages
|
|
if (messages.length == 1) {
|
|
if ($settings.saveChatHistory ?? true) {
|
|
chat = await createNewChat(localStorage.token, {
|
|
id: $chatId,
|
|
title: 'New Chat',
|
|
models: selectedModels,
|
|
system: $settings.system ?? undefined,
|
|
options: {
|
|
...($settings.options ?? {})
|
|
},
|
|
messages: messages,
|
|
history: history,
|
|
timestamp: Date.now()
|
|
});
|
|
await chats.set(await getChatList(localStorage.token));
|
|
await chatId.set(chat.id);
|
|
} else {
|
|
await chatId.set('local');
|
|
}
|
|
await tick();
|
|
}
|
|
|
|
// Reset chat input textarea
|
|
prompt = '';
|
|
files = [];
|
|
|
|
// Send prompt
|
|
await sendPrompt(userPrompt, userMessageId);
|
|
}
|
|
};
|
|
|
|
const sendPrompt = async (prompt, parentId) => {
|
|
const _chatId = JSON.parse(JSON.stringify($chatId));
|
|
|
|
const docs = messages
|
|
.filter((message) => message?.files ?? null)
|
|
.map((message) => message.files.filter((item) => item.type === 'doc'))
|
|
.flat(1);
|
|
|
|
console.log(docs);
|
|
if (docs.length > 0) {
|
|
processing = 'Reading';
|
|
const query = history.messages[parentId].content;
|
|
|
|
let relevantContexts = await Promise.all(
|
|
docs.map(async (doc) => {
|
|
return await queryVectorDB(localStorage.token, doc.collection_name, query, 4).catch(
|
|
(error) => {
|
|
console.log(error);
|
|
return null;
|
|
}
|
|
);
|
|
})
|
|
);
|
|
relevantContexts = relevantContexts.filter((context) => context);
|
|
|
|
const contextString = relevantContexts.reduce((a, context, i, arr) => {
|
|
return `${a}${context.documents.join(' ')}\n`;
|
|
}, '');
|
|
|
|
console.log(contextString);
|
|
|
|
history.messages[parentId].raContent = RAGTemplate(contextString, query);
|
|
history.messages[parentId].contexts = relevantContexts;
|
|
await tick();
|
|
processing = '';
|
|
}
|
|
|
|
await Promise.all(
|
|
selectedModels.map(async (model) => {
|
|
console.log(model);
|
|
const modelTag = $models.filter((m) => m.name === model).at(0);
|
|
|
|
if (modelTag?.external) {
|
|
await sendPromptOpenAI(model, prompt, parentId, _chatId);
|
|
} else if (modelTag) {
|
|
await sendPromptOllama(model, prompt, parentId, _chatId);
|
|
} else {
|
|
toast.error(`Model ${model} not found`);
|
|
}
|
|
})
|
|
);
|
|
|
|
await chats.set(await getChatList(localStorage.token));
|
|
};
|
|
|
|
const sendPromptOllama = async (model, userPrompt, parentId, _chatId) => {
|
|
// Create response message
|
|
let responseMessageId = uuidv4();
|
|
let responseMessage = {
|
|
parentId: parentId,
|
|
id: responseMessageId,
|
|
childrenIds: [],
|
|
role: 'assistant',
|
|
content: '',
|
|
model: model,
|
|
timestamp: Math.floor(Date.now() / 1000) // Unix epoch
|
|
};
|
|
|
|
// Add message to history and Set currentId to messageId
|
|
history.messages[responseMessageId] = responseMessage;
|
|
history.currentId = responseMessageId;
|
|
|
|
// Append messageId to childrenIds of parent message
|
|
if (parentId !== null) {
|
|
history.messages[parentId].childrenIds = [
|
|
...history.messages[parentId].childrenIds,
|
|
responseMessageId
|
|
];
|
|
}
|
|
|
|
// Wait until history/message have been updated
|
|
await tick();
|
|
|
|
// Scroll down
|
|
window.scrollTo({ top: document.body.scrollHeight });
|
|
|
|
const res = await generateChatCompletion(localStorage.token, {
|
|
model: model,
|
|
messages: [
|
|
$settings.system
|
|
? {
|
|
role: 'system',
|
|
content: $settings.system
|
|
}
|
|
: undefined,
|
|
...messages
|
|
]
|
|
.filter((message) => message)
|
|
.map((message, idx, arr) => ({
|
|
role: message.role,
|
|
content: arr.length - 2 !== idx ? message.content : message?.raContent ?? message.content,
|
|
...(message.files && {
|
|
images: message.files
|
|
.filter((file) => file.type === 'image')
|
|
.map((file) => file.url.slice(file.url.indexOf(',') + 1))
|
|
})
|
|
})),
|
|
options: {
|
|
...($settings.options ?? {})
|
|
},
|
|
format: $settings.requestFormat ?? undefined
|
|
});
|
|
|
|
if (res && res.ok) {
|
|
const reader = res.body
|
|
.pipeThrough(new TextDecoderStream())
|
|
.pipeThrough(splitStream('\n'))
|
|
.getReader();
|
|
|
|
while (true) {
|
|
const { value, done } = await reader.read();
|
|
if (done || stopResponseFlag || _chatId !== $chatId) {
|
|
responseMessage.done = true;
|
|
messages = messages;
|
|
break;
|
|
}
|
|
|
|
try {
|
|
let lines = value.split('\n');
|
|
|
|
for (const line of lines) {
|
|
if (line !== '') {
|
|
console.log(line);
|
|
let data = JSON.parse(line);
|
|
|
|
if ('detail' in data) {
|
|
throw data;
|
|
}
|
|
|
|
if (data.done == false) {
|
|
if (responseMessage.content == '' && data.message.content == '\n') {
|
|
continue;
|
|
} else {
|
|
responseMessage.content += data.message.content;
|
|
messages = messages;
|
|
}
|
|
} else {
|
|
responseMessage.done = true;
|
|
|
|
if (responseMessage.content == '') {
|
|
responseMessage.error = true;
|
|
responseMessage.content =
|
|
'Oops! No text generated from Ollama, Please try again.';
|
|
}
|
|
|
|
responseMessage.context = data.context ?? null;
|
|
responseMessage.info = {
|
|
total_duration: data.total_duration,
|
|
load_duration: data.load_duration,
|
|
sample_count: data.sample_count,
|
|
sample_duration: data.sample_duration,
|
|
prompt_eval_count: data.prompt_eval_count,
|
|
prompt_eval_duration: data.prompt_eval_duration,
|
|
eval_count: data.eval_count,
|
|
eval_duration: data.eval_duration
|
|
};
|
|
messages = messages;
|
|
|
|
if ($settings.notificationEnabled && !document.hasFocus()) {
|
|
const notification = new Notification(
|
|
selectedModelfile
|
|
? `${
|
|
selectedModelfile.title.charAt(0).toUpperCase() +
|
|
selectedModelfile.title.slice(1)
|
|
}`
|
|
: `Ollama - ${model}`,
|
|
{
|
|
body: responseMessage.content,
|
|
icon: selectedModelfile?.imageUrl ?? '/favicon.png'
|
|
}
|
|
);
|
|
}
|
|
|
|
if ($settings.responseAutoCopy) {
|
|
copyToClipboard(responseMessage.content);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
} catch (error) {
|
|
console.log(error);
|
|
if ('detail' in error) {
|
|
toast.error(error.detail);
|
|
}
|
|
break;
|
|
}
|
|
|
|
if (autoScroll) {
|
|
window.scrollTo({ top: document.body.scrollHeight });
|
|
}
|
|
}
|
|
|
|
if ($chatId == _chatId) {
|
|
if ($settings.saveChatHistory ?? true) {
|
|
chat = await updateChatById(localStorage.token, _chatId, {
|
|
messages: messages,
|
|
history: history
|
|
});
|
|
await chats.set(await getChatList(localStorage.token));
|
|
}
|
|
}
|
|
} else {
|
|
if (res !== null) {
|
|
const error = await res.json();
|
|
console.log(error);
|
|
if ('detail' in error) {
|
|
toast.error(error.detail);
|
|
responseMessage.content = error.detail;
|
|
} else {
|
|
toast.error(error.error);
|
|
responseMessage.content = error.error;
|
|
}
|
|
} else {
|
|
toast.error(`Uh-oh! There was an issue connecting to Ollama.`);
|
|
responseMessage.content = `Uh-oh! There was an issue connecting to Ollama.`;
|
|
}
|
|
|
|
responseMessage.error = true;
|
|
responseMessage.content = `Uh-oh! There was an issue connecting to Ollama.`;
|
|
responseMessage.done = true;
|
|
messages = messages;
|
|
}
|
|
|
|
stopResponseFlag = false;
|
|
await tick();
|
|
|
|
if (autoScroll) {
|
|
window.scrollTo({ top: document.body.scrollHeight });
|
|
}
|
|
|
|
if (messages.length == 2 && messages.at(1).content !== '') {
|
|
window.history.replaceState(history.state, '', `/c/${_chatId}`);
|
|
await generateChatTitle(_chatId, userPrompt);
|
|
}
|
|
};
|
|
|
|
const sendPromptOpenAI = async (model, userPrompt, parentId, _chatId) => {
|
|
let responseMessageId = uuidv4();
|
|
|
|
let responseMessage = {
|
|
parentId: parentId,
|
|
id: responseMessageId,
|
|
childrenIds: [],
|
|
role: 'assistant',
|
|
content: '',
|
|
model: model,
|
|
timestamp: Math.floor(Date.now() / 1000) // Unix epoch
|
|
};
|
|
|
|
history.messages[responseMessageId] = responseMessage;
|
|
history.currentId = responseMessageId;
|
|
if (parentId !== null) {
|
|
history.messages[parentId].childrenIds = [
|
|
...history.messages[parentId].childrenIds,
|
|
responseMessageId
|
|
];
|
|
}
|
|
|
|
window.scrollTo({ top: document.body.scrollHeight });
|
|
|
|
const res = await generateOpenAIChatCompletion(localStorage.token, {
|
|
model: model,
|
|
stream: true,
|
|
messages: [
|
|
$settings.system
|
|
? {
|
|
role: 'system',
|
|
content: $settings.system
|
|
}
|
|
: undefined,
|
|
...messages
|
|
]
|
|
.filter((message) => message)
|
|
.map((message, idx, arr) => ({
|
|
role: message.role,
|
|
...(message.files
|
|
? {
|
|
content: [
|
|
{
|
|
type: 'text',
|
|
text:
|
|
arr.length - 1 !== idx
|
|
? message.content
|
|
: message?.raContent ?? message.content
|
|
},
|
|
...message.files
|
|
.filter((file) => file.type === 'image')
|
|
.map((file) => ({
|
|
type: 'image_url',
|
|
image_url: {
|
|
url: file.url
|
|
}
|
|
}))
|
|
]
|
|
}
|
|
: {
|
|
content:
|
|
arr.length - 1 !== idx ? message.content : message?.raContent ?? message.content
|
|
})
|
|
})),
|
|
seed: $settings?.options?.seed ?? undefined,
|
|
stop: $settings?.options?.stop ?? undefined,
|
|
temperature: $settings?.options?.temperature ?? undefined,
|
|
top_p: $settings?.options?.top_p ?? undefined,
|
|
num_ctx: $settings?.options?.num_ctx ?? undefined,
|
|
frequency_penalty: $settings?.options?.repeat_penalty ?? undefined,
|
|
max_tokens: $settings?.options?.num_predict ?? undefined
|
|
});
|
|
|
|
if (res && res.ok) {
|
|
const reader = res.body
|
|
.pipeThrough(new TextDecoderStream())
|
|
.pipeThrough(splitStream('\n'))
|
|
.getReader();
|
|
|
|
while (true) {
|
|
const { value, done } = await reader.read();
|
|
if (done || stopResponseFlag || _chatId !== $chatId) {
|
|
responseMessage.done = true;
|
|
messages = messages;
|
|
break;
|
|
}
|
|
|
|
try {
|
|
let lines = value.split('\n');
|
|
|
|
for (const line of lines) {
|
|
if (line !== '') {
|
|
console.log(line);
|
|
if (line === 'data: [DONE]') {
|
|
responseMessage.done = true;
|
|
messages = messages;
|
|
} else {
|
|
let data = JSON.parse(line.replace(/^data: /, ''));
|
|
console.log(data);
|
|
|
|
if (responseMessage.content == '' && data.choices[0].delta.content == '\n') {
|
|
continue;
|
|
} else {
|
|
responseMessage.content += data.choices[0].delta.content ?? '';
|
|
messages = messages;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
} catch (error) {
|
|
console.log(error);
|
|
}
|
|
|
|
if ($settings.notificationEnabled && !document.hasFocus()) {
|
|
const notification = new Notification(`OpenAI ${model}`, {
|
|
body: responseMessage.content,
|
|
icon: '/favicon.png'
|
|
});
|
|
}
|
|
|
|
if ($settings.responseAutoCopy) {
|
|
copyToClipboard(responseMessage.content);
|
|
}
|
|
|
|
if (autoScroll) {
|
|
window.scrollTo({ top: document.body.scrollHeight });
|
|
}
|
|
}
|
|
|
|
if ($chatId == _chatId) {
|
|
if ($settings.saveChatHistory ?? true) {
|
|
chat = await updateChatById(localStorage.token, _chatId, {
|
|
messages: messages,
|
|
history: history
|
|
});
|
|
await chats.set(await getChatList(localStorage.token));
|
|
}
|
|
}
|
|
} else {
|
|
if (res !== null) {
|
|
const error = await res.json();
|
|
console.log(error);
|
|
if ('detail' in error) {
|
|
toast.error(error.detail);
|
|
responseMessage.content = error.detail;
|
|
} else {
|
|
if ('message' in error.error) {
|
|
toast.error(error.error.message);
|
|
responseMessage.content = error.error.message;
|
|
} else {
|
|
toast.error(error.error);
|
|
responseMessage.content = error.error;
|
|
}
|
|
}
|
|
} else {
|
|
toast.error(`Uh-oh! There was an issue connecting to ${model}.`);
|
|
responseMessage.content = `Uh-oh! There was an issue connecting to ${model}.`;
|
|
}
|
|
|
|
responseMessage.error = true;
|
|
responseMessage.content = `Uh-oh! There was an issue connecting to ${model}.`;
|
|
responseMessage.done = true;
|
|
messages = messages;
|
|
}
|
|
|
|
stopResponseFlag = false;
|
|
await tick();
|
|
|
|
if (autoScroll) {
|
|
window.scrollTo({ top: document.body.scrollHeight });
|
|
}
|
|
|
|
if (messages.length == 2) {
|
|
window.history.replaceState(history.state, '', `/c/${_chatId}`);
|
|
await setChatTitle(_chatId, userPrompt);
|
|
}
|
|
};
|
|
|
|
const stopResponse = () => {
|
|
stopResponseFlag = true;
|
|
console.log('stopResponse');
|
|
};
|
|
|
|
const regenerateResponse = async () => {
|
|
console.log('regenerateResponse');
|
|
if (messages.length != 0 && messages.at(-1).done == true) {
|
|
messages.splice(messages.length - 1, 1);
|
|
messages = messages;
|
|
|
|
let userMessage = messages.at(-1);
|
|
let userPrompt = userMessage.content;
|
|
|
|
await sendPrompt(userPrompt, userMessage.id);
|
|
}
|
|
};
|
|
|
|
const generateChatTitle = async (_chatId, userPrompt) => {
|
|
if ($settings.titleAutoGenerate ?? true) {
|
|
const title = await generateTitle(
|
|
localStorage.token,
|
|
$settings?.titleAutoGenerateModel ?? selectedModels[0],
|
|
userPrompt
|
|
);
|
|
|
|
if (title) {
|
|
await setChatTitle(_chatId, title);
|
|
}
|
|
} else {
|
|
await setChatTitle(_chatId, `${userPrompt}`);
|
|
}
|
|
};
|
|
|
|
const setChatTitle = async (_chatId, _title) => {
|
|
if (_chatId === $chatId) {
|
|
title = _title;
|
|
}
|
|
|
|
if ($settings.saveChatHistory ?? true) {
|
|
chat = await updateChatById(localStorage.token, _chatId, { title: _title });
|
|
await chats.set(await getChatList(localStorage.token));
|
|
}
|
|
};
|
|
</script>
|
|
|
|
<svelte:window
|
|
on:scroll={(e) => {
|
|
autoScroll = window.innerHeight + window.scrollY >= document.body.offsetHeight - 40;
|
|
}}
|
|
/>
|
|
|
|
<Navbar {title} shareEnabled={messages.length > 0} {initNewChat} />
|
|
<div class="min-h-screen w-full flex justify-center">
|
|
<div class=" py-2.5 flex flex-col justify-between w-full">
|
|
<div class="max-w-2xl mx-auto w-full px-3 md:px-0 mt-10">
|
|
<ModelSelector bind:selectedModels disabled={messages.length > 0} />
|
|
</div>
|
|
|
|
<div class=" h-full mt-10 mb-32 w-full flex flex-col">
|
|
<Messages
|
|
chatId={$chatId}
|
|
{selectedModels}
|
|
{selectedModelfiles}
|
|
{processing}
|
|
bind:history
|
|
bind:messages
|
|
bind:autoScroll
|
|
bottomPadding={files.length > 0}
|
|
{sendPrompt}
|
|
{regenerateResponse}
|
|
/>
|
|
</div>
|
|
</div>
|
|
|
|
<MessageInput
|
|
bind:files
|
|
bind:prompt
|
|
bind:autoScroll
|
|
suggestionPrompts={selectedModelfile?.suggestionPrompts ?? [
|
|
{
|
|
title: ['Help me study', 'vocabulary for a college entrance exam'],
|
|
content: `Help me study vocabulary: write a sentence for me to fill in the blank, and I'll try to pick the correct option.`
|
|
},
|
|
{
|
|
title: ['Give me ideas', `for what to do with my kids' art`],
|
|
content: `What are 5 creative things I could do with my kids' art? I don't want to throw them away, but it's also so much clutter.`
|
|
},
|
|
{
|
|
title: ['Tell me a fun fact', 'about the Roman Empire'],
|
|
content: 'Tell me a random fun fact about the Roman Empire'
|
|
},
|
|
{
|
|
title: ['Show me a code snippet', `of a website's sticky header`],
|
|
content: `Show me a code snippet of a website's sticky header in CSS and JavaScript.`
|
|
}
|
|
]}
|
|
{messages}
|
|
{submitPrompt}
|
|
{stopResponse}
|
|
/>
|
|
</div>
|