123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863 |
- <script lang="ts">
- import { v4 as uuidv4 } from 'uuid';
- import toast from 'svelte-french-toast';
- import { onMount, tick } from 'svelte';
- import { goto } from '$app/navigation';
- import { page } from '$app/stores';
- import {
- models,
- modelfiles,
- user,
- settings,
- chats,
- chatId,
- config,
- tags as _tags
- } from '$lib/stores';
- import { copyToClipboard, splitStream } from '$lib/utils';
- import { generateChatCompletion, cancelChatCompletion, generateTitle } from '$lib/apis/ollama';
- import {
- addTagById,
- createNewChat,
- deleteTagById,
- getAllChatTags,
- getChatList,
- getTagsById,
- updateChatById
- } from '$lib/apis/chats';
- import { queryCollection, queryDoc } from '$lib/apis/rag';
- import { generateOpenAIChatCompletion } from '$lib/apis/openai';
- import MessageInput from '$lib/components/chat/MessageInput.svelte';
- import Messages from '$lib/components/chat/Messages.svelte';
- import ModelSelector from '$lib/components/chat/ModelSelector.svelte';
- import Navbar from '$lib/components/layout/Navbar.svelte';
- import { RAGTemplate } from '$lib/utils/rag';
- import { LITELLM_API_BASE_URL, OPENAI_API_BASE_URL } from '$lib/constants';
- import { WEBUI_BASE_URL } from '$lib/constants';
- let stopResponseFlag = false;
- let autoScroll = true;
- let processing = '';
- let currentRequestId = null;
- let selectedModels = [''];
- let selectedModelfile = null;
- $: selectedModelfile =
- selectedModels.length === 1 &&
- $modelfiles.filter((modelfile) => modelfile.tagName === selectedModels[0]).length > 0
- ? $modelfiles.filter((modelfile) => modelfile.tagName === selectedModels[0])[0]
- : null;
- let selectedModelfiles = {};
- $: selectedModelfiles = selectedModels.reduce((a, tagName, i, arr) => {
- const modelfile =
- $modelfiles.filter((modelfile) => modelfile.tagName === tagName)?.at(0) ?? undefined;
- return {
- ...a,
- ...(modelfile && { [tagName]: modelfile })
- };
- }, {});
- let chat = null;
- let tags = [];
- let title = '';
- let prompt = '';
- let files = [];
- let messages = [];
- let history = {
- messages: {},
- currentId: null
- };
- $: if (history.currentId !== null) {
- let _messages = [];
- let currentMessage = history.messages[history.currentId];
- while (currentMessage !== null) {
- _messages.unshift({ ...currentMessage });
- currentMessage =
- currentMessage.parentId !== null ? history.messages[currentMessage.parentId] : null;
- }
- messages = _messages;
- } else {
- messages = [];
- }
- onMount(async () => {
- await initNewChat();
- });
- //////////////////////////
- // Web functions
- //////////////////////////
- const initNewChat = async () => {
- if (currentRequestId !== null) {
- await cancelChatCompletion(localStorage.token, currentRequestId);
- currentRequestId = null;
- }
- window.history.replaceState(history.state, '', `/`);
- console.log('initNewChat');
- await chatId.set('');
- console.log($chatId);
- autoScroll = true;
- title = '';
- messages = [];
- history = {
- messages: {},
- currentId: null
- };
- console.log($config);
- if ($page.url.searchParams.get('models')) {
- selectedModels = $page.url.searchParams.get('models')?.split(',');
- } else if ($settings?.models) {
- selectedModels = $settings?.models;
- } else if ($config?.default_models) {
- selectedModels = $config?.default_models.split(',');
- } else {
- selectedModels = [''];
- }
- selectedModels = selectedModels.map((modelId) =>
- $models.map((m) => m.id).includes(modelId) ? modelId : ''
- );
- let _settings = JSON.parse(localStorage.getItem('settings') ?? '{}');
- settings.set({
- ..._settings
- });
- };
- const scrollToBottom = () => {
- const element = document.getElementById('messages-container');
- element.scrollTop = element.scrollHeight;
- };
- //////////////////////////
- // Ollama functions
- //////////////////////////
- const submitPrompt = async (userPrompt, _user = null) => {
- console.log('submitPrompt', $chatId);
- selectedModels = selectedModels.map((modelId) =>
- $models.map((m) => m.id).includes(modelId) ? modelId : ''
- );
- if (selectedModels.includes('')) {
- toast.error('Model not selected');
- } else if (messages.length != 0 && messages.at(-1).done != true) {
- // Response not done
- console.log('wait');
- } else if (
- files.length > 0 &&
- files.filter((file) => file.upload_status === false).length > 0
- ) {
- // Upload not done
- toast.error(
- `Oops! Hold tight! Your files are still in the processing oven. We're cooking them up to perfection. Please be patient and we'll let you know once they're ready.`
- );
- } else {
- // Reset chat message textarea height
- document.getElementById('chat-textarea').style.height = '';
- // Create user message
- let userMessageId = uuidv4();
- let userMessage = {
- id: userMessageId,
- parentId: messages.length !== 0 ? messages.at(-1).id : null,
- childrenIds: [],
- role: 'user',
- user: _user ?? undefined,
- content: userPrompt,
- files: files.length > 0 ? files : undefined,
- timestamp: Math.floor(Date.now() / 1000) // Unix epoch
- };
- // Add message to history and Set currentId to messageId
- history.messages[userMessageId] = userMessage;
- history.currentId = userMessageId;
- // Append messageId to childrenIds of parent message
- if (messages.length !== 0) {
- history.messages[messages.at(-1).id].childrenIds.push(userMessageId);
- }
- // Wait until history/message have been updated
- await tick();
- // Create new chat if only one message in messages
- if (messages.length == 1) {
- if ($settings.saveChatHistory ?? true) {
- chat = await createNewChat(localStorage.token, {
- id: $chatId,
- title: 'New Chat',
- models: selectedModels,
- system: $settings.system ?? undefined,
- options: {
- ...($settings.options ?? {})
- },
- messages: messages,
- history: history,
- tags: [],
- timestamp: Date.now()
- });
- await chats.set(await getChatList(localStorage.token));
- await chatId.set(chat.id);
- } else {
- await chatId.set('local');
- }
- await tick();
- }
- // Reset chat input textarea
- prompt = '';
- files = [];
- // Send prompt
- await sendPrompt(userPrompt, userMessageId);
- }
- };
- const sendPrompt = async (prompt, parentId) => {
- const _chatId = JSON.parse(JSON.stringify($chatId));
- const docs = messages
- .filter((message) => message?.files ?? null)
- .map((message) =>
- message.files.filter((item) => item.type === 'doc' || item.type === 'collection')
- )
- .flat(1);
- console.log(docs);
- if (docs.length > 0) {
- processing = 'Reading';
- const query = history.messages[parentId].content;
- let relevantContexts = await Promise.all(
- docs.map(async (doc) => {
- if (doc.type === 'collection') {
- return await queryCollection(localStorage.token, doc.collection_names, query, 4).catch(
- (error) => {
- console.log(error);
- return null;
- }
- );
- } else {
- return await queryDoc(localStorage.token, doc.collection_name, query, 4).catch(
- (error) => {
- console.log(error);
- return null;
- }
- );
- }
- })
- );
- relevantContexts = relevantContexts.filter((context) => context);
- const contextString = relevantContexts.reduce((a, context, i, arr) => {
- return `${a}${context.documents.join(' ')}\n`;
- }, '');
- console.log(contextString);
- history.messages[parentId].raContent = await RAGTemplate(
- localStorage.token,
- contextString,
- query
- );
- history.messages[parentId].contexts = relevantContexts;
- await tick();
- processing = '';
- }
- await Promise.all(
- selectedModels.map(async (modelId) => {
- const model = $models.filter((m) => m.id === modelId).at(0);
- if (model) {
- // Create response message
- let responseMessageId = uuidv4();
- let responseMessage = {
- parentId: parentId,
- id: responseMessageId,
- childrenIds: [],
- role: 'assistant',
- content: '',
- model: model.id,
- timestamp: Math.floor(Date.now() / 1000) // Unix epoch
- };
- // Add message to history and Set currentId to messageId
- history.messages[responseMessageId] = responseMessage;
- history.currentId = responseMessageId;
- // Append messageId to childrenIds of parent message
- if (parentId !== null) {
- history.messages[parentId].childrenIds = [
- ...history.messages[parentId].childrenIds,
- responseMessageId
- ];
- }
- if (model?.external) {
- await sendPromptOpenAI(model, prompt, responseMessageId, _chatId);
- } else if (model) {
- await sendPromptOllama(model, prompt, responseMessageId, _chatId);
- }
- } else {
- toast.error(`Model ${modelId} not found`);
- }
- })
- );
- await chats.set(await getChatList(localStorage.token));
- };
- const sendPromptOllama = async (model, userPrompt, responseMessageId, _chatId) => {
- model = model.id;
- const responseMessage = history.messages[responseMessageId];
- // Wait until history/message have been updated
- await tick();
- // Scroll down
- scrollToBottom();
- const messagesBody = [
- $settings.system
- ? {
- role: 'system',
- content: $settings.system
- }
- : undefined,
- ...messages.filter((message) => !message.deleted)
- ]
- .filter((message) => message)
- .map((message, idx, arr) => ({
- role: message.role,
- content: arr.length - 2 !== idx ? message.content : message?.raContent ?? message.content,
- ...(message.files && {
- images: message.files
- .filter((file) => file.type === 'image')
- .map((file) => file.url.slice(file.url.indexOf(',') + 1))
- })
- }));
- let lastImageIndex = -1;
- // Find the index of the last object with images
- messagesBody.forEach((item, index) => {
- if (item.images) {
- lastImageIndex = index;
- }
- });
- // Remove images from all but the last one
- messagesBody.forEach((item, index) => {
- if (index !== lastImageIndex) {
- delete item.images;
- }
- });
- const [res, controller] = await generateChatCompletion(localStorage.token, {
- model: model,
- messages: messagesBody,
- options: {
- ...($settings.options ?? {})
- },
- format: $settings.requestFormat ?? undefined,
- keep_alive: $settings.keepAlive ?? undefined
- });
- if (res && res.ok) {
- console.log('controller', controller);
- const reader = res.body
- .pipeThrough(new TextDecoderStream())
- .pipeThrough(splitStream('\n'))
- .getReader();
- while (true) {
- const { value, done } = await reader.read();
- if (done || stopResponseFlag || _chatId !== $chatId) {
- responseMessage.done = true;
- messages = messages;
- if (stopResponseFlag) {
- controller.abort('User: Stop Response');
- await cancelChatCompletion(localStorage.token, currentRequestId);
- }
- currentRequestId = null;
- break;
- }
- try {
- let lines = value.split('\n');
- for (const line of lines) {
- if (line !== '') {
- console.log(line);
- let data = JSON.parse(line);
- if ('detail' in data) {
- throw data;
- }
- if ('id' in data) {
- console.log(data);
- currentRequestId = data.id;
- } else {
- if (data.done == false) {
- if (responseMessage.content == '' && data.message.content == '\n') {
- continue;
- } else {
- responseMessage.content += data.message.content;
- messages = messages;
- }
- } else {
- responseMessage.done = true;
- if (responseMessage.content == '') {
- responseMessage.error = true;
- responseMessage.content =
- 'Oops! No text generated from Ollama, Please try again.';
- }
- responseMessage.context = data.context ?? null;
- responseMessage.info = {
- total_duration: data.total_duration,
- load_duration: data.load_duration,
- sample_count: data.sample_count,
- sample_duration: data.sample_duration,
- prompt_eval_count: data.prompt_eval_count,
- prompt_eval_duration: data.prompt_eval_duration,
- eval_count: data.eval_count,
- eval_duration: data.eval_duration
- };
- messages = messages;
- if ($settings.notificationEnabled && !document.hasFocus()) {
- const notification = new Notification(
- selectedModelfile
- ? `${
- selectedModelfile.title.charAt(0).toUpperCase() +
- selectedModelfile.title.slice(1)
- }`
- : `${model}`,
- {
- body: responseMessage.content,
- icon: selectedModelfile?.imageUrl ?? `${WEBUI_BASE_URL}/static/favicon.png`
- }
- );
- }
- if ($settings.responseAutoCopy) {
- copyToClipboard(responseMessage.content);
- }
- if ($settings.responseAutoPlayback) {
- await tick();
- document.getElementById(`speak-button-${responseMessage.id}`)?.click();
- }
- }
- }
- }
- }
- } catch (error) {
- console.log(error);
- if ('detail' in error) {
- toast.error(error.detail);
- }
- break;
- }
- if (autoScroll) {
- scrollToBottom();
- }
- }
- if ($chatId == _chatId) {
- if ($settings.saveChatHistory ?? true) {
- chat = await updateChatById(localStorage.token, _chatId, {
- messages: messages,
- history: history
- });
- await chats.set(await getChatList(localStorage.token));
- }
- }
- } else {
- if (res !== null) {
- const error = await res.json();
- console.log(error);
- if ('detail' in error) {
- toast.error(error.detail);
- responseMessage.content = error.detail;
- } else {
- toast.error(error.error);
- responseMessage.content = error.error;
- }
- } else {
- toast.error(`Uh-oh! There was an issue connecting to Ollama.`);
- responseMessage.content = `Uh-oh! There was an issue connecting to Ollama.`;
- }
- responseMessage.error = true;
- responseMessage.content = `Uh-oh! There was an issue connecting to Ollama.`;
- responseMessage.done = true;
- messages = messages;
- }
- stopResponseFlag = false;
- await tick();
- if (autoScroll) {
- scrollToBottom();
- }
- if (messages.length == 2 && messages.at(1).content !== '') {
- window.history.replaceState(history.state, '', `/c/${_chatId}`);
- await generateChatTitle(_chatId, userPrompt);
- }
- };
- const sendPromptOpenAI = async (model, userPrompt, responseMessageId, _chatId) => {
- const responseMessage = history.messages[responseMessageId];
- scrollToBottom();
- const res = await generateOpenAIChatCompletion(
- localStorage.token,
- {
- model: model.id,
- stream: true,
- messages: [
- $settings.system
- ? {
- role: 'system',
- content: $settings.system
- }
- : undefined,
- ...messages.filter((message) => !message.deleted)
- ]
- .filter((message) => message)
- .map((message, idx, arr) => ({
- role: message.role,
- ...(message.files?.filter((file) => file.type === 'image').length > 0 ?? false
- ? {
- content: [
- {
- type: 'text',
- text:
- arr.length - 1 !== idx
- ? message.content
- : message?.raContent ?? message.content
- },
- ...message.files
- .filter((file) => file.type === 'image')
- .map((file) => ({
- type: 'image_url',
- image_url: {
- url: file.url
- }
- }))
- ]
- }
- : {
- content:
- arr.length - 1 !== idx ? message.content : message?.raContent ?? message.content
- })
- })),
- seed: $settings?.options?.seed ?? undefined,
- stop: $settings?.options?.stop ?? undefined,
- temperature: $settings?.options?.temperature ?? undefined,
- top_p: $settings?.options?.top_p ?? undefined,
- num_ctx: $settings?.options?.num_ctx ?? undefined,
- frequency_penalty: $settings?.options?.repeat_penalty ?? undefined,
- max_tokens: $settings?.options?.num_predict ?? undefined
- },
- model.source === 'litellm' ? `${LITELLM_API_BASE_URL}/v1` : `${OPENAI_API_BASE_URL}`
- );
- if (res && res.ok) {
- const reader = res.body
- .pipeThrough(new TextDecoderStream())
- .pipeThrough(splitStream('\n'))
- .getReader();
- while (true) {
- const { value, done } = await reader.read();
- if (done || stopResponseFlag || _chatId !== $chatId) {
- responseMessage.done = true;
- messages = messages;
- break;
- }
- try {
- let lines = value.split('\n');
- for (const line of lines) {
- if (line !== '') {
- console.log(line);
- if (line === 'data: [DONE]') {
- responseMessage.done = true;
- messages = messages;
- } else {
- let data = JSON.parse(line.replace(/^data: /, ''));
- console.log(data);
- if (responseMessage.content == '' && data.choices[0].delta.content == '\n') {
- continue;
- } else {
- responseMessage.content += data.choices[0].delta.content ?? '';
- messages = messages;
- }
- }
- }
- }
- } catch (error) {
- console.log(error);
- }
- if ($settings.notificationEnabled && !document.hasFocus()) {
- const notification = new Notification(`OpenAI ${model}`, {
- body: responseMessage.content,
- icon: `${WEBUI_BASE_URL}/static/favicon.png`
- });
- }
- if ($settings.responseAutoCopy) {
- copyToClipboard(responseMessage.content);
- }
- if ($settings.responseAutoPlayback) {
- await tick();
- document.getElementById(`speak-button-${responseMessage.id}`)?.click();
- }
- if (autoScroll) {
- scrollToBottom();
- }
- }
- if ($chatId == _chatId) {
- if ($settings.saveChatHistory ?? true) {
- chat = await updateChatById(localStorage.token, _chatId, {
- messages: messages,
- history: history
- });
- await chats.set(await getChatList(localStorage.token));
- }
- }
- } else {
- if (res !== null) {
- const error = await res.json();
- console.log(error);
- if ('detail' in error) {
- toast.error(error.detail);
- responseMessage.content = error.detail;
- } else {
- if ('message' in error.error) {
- toast.error(error.error.message);
- responseMessage.content = error.error.message;
- } else {
- toast.error(error.error);
- responseMessage.content = error.error;
- }
- }
- } else {
- toast.error(`Uh-oh! There was an issue connecting to ${model}.`);
- responseMessage.content = `Uh-oh! There was an issue connecting to ${model}.`;
- }
- responseMessage.error = true;
- responseMessage.content = `Uh-oh! There was an issue connecting to ${model}.`;
- responseMessage.done = true;
- messages = messages;
- }
- stopResponseFlag = false;
- await tick();
- if (autoScroll) {
- scrollToBottom();
- }
- if (messages.length == 2) {
- window.history.replaceState(history.state, '', `/c/${_chatId}`);
- await setChatTitle(_chatId, userPrompt);
- }
- };
- const stopResponse = () => {
- stopResponseFlag = true;
- console.log('stopResponse');
- };
- const regenerateResponse = async () => {
- console.log('regenerateResponse');
- if (messages.length != 0 && messages.at(-1).done == true) {
- messages.splice(messages.length - 1, 1);
- messages = messages;
- let userMessage = messages.at(-1);
- let userPrompt = userMessage.content;
- await sendPrompt(userPrompt, userMessage.id);
- }
- };
- const continueGeneration = async () => {
- console.log('continueGeneration');
- const _chatId = JSON.parse(JSON.stringify($chatId));
- if (messages.length != 0 && messages.at(-1).done == true) {
- const responseMessage = history.messages[history.currentId];
- responseMessage.done = false;
- await tick();
- const modelTag = $models.filter((m) => m.name === responseMessage.model).at(0);
- if (modelTag?.external) {
- await sendPromptOpenAI(
- responseMessage.model,
- history.messages[responseMessage.parentId].content,
- responseMessage.id,
- _chatId
- );
- } else if (modelTag) {
- await sendPromptOllama(
- responseMessage.model,
- history.messages[responseMessage.parentId].content,
- responseMessage.id,
- _chatId
- );
- } else {
- toast.error(`Model ${model} not found`);
- }
- }
- };
- const generateChatTitle = async (_chatId, userPrompt) => {
- if ($settings.titleAutoGenerate ?? true) {
- const title = await generateTitle(
- localStorage.token,
- $settings?.titleGenerationPrompt ??
- "Create a concise, 3-5 word phrase as a header for the following query, strictly adhering to the 3-5 word limit and avoiding the use of the word 'title': {{prompt}}",
- $settings?.titleAutoGenerateModel ?? selectedModels[0],
- userPrompt
- );
- if (title) {
- await setChatTitle(_chatId, title);
- }
- } else {
- await setChatTitle(_chatId, `${userPrompt}`);
- }
- };
- const getTags = async () => {
- return await getTagsById(localStorage.token, $chatId).catch(async (error) => {
- return [];
- });
- };
- const addTag = async (tagName) => {
- const res = await addTagById(localStorage.token, $chatId, tagName);
- tags = await getTags();
- chat = await updateChatById(localStorage.token, $chatId, {
- tags: tags
- });
- _tags.set(await getAllChatTags(localStorage.token));
- };
- const deleteTag = async (tagName) => {
- const res = await deleteTagById(localStorage.token, $chatId, tagName);
- tags = await getTags();
- chat = await updateChatById(localStorage.token, $chatId, {
- tags: tags
- });
- _tags.set(await getAllChatTags(localStorage.token));
- };
- const setChatTitle = async (_chatId, _title) => {
- if (_chatId === $chatId) {
- title = _title;
- }
- if ($settings.saveChatHistory ?? true) {
- chat = await updateChatById(localStorage.token, _chatId, { title: _title });
- await chats.set(await getChatList(localStorage.token));
- }
- };
- </script>
- <div class="h-screen max-h-[100dvh] w-full flex flex-col">
- <Navbar {title} shareEnabled={messages.length > 0} {initNewChat} {tags} {addTag} {deleteTag} />
- <div class="flex flex-col flex-auto">
- <div
- class=" pb-2.5 flex flex-col justify-between w-full flex-auto overflow-auto h-0"
- id="messages-container"
- on:scroll={(e) => {
- autoScroll = e.target.scrollHeight - e.target.scrollTop <= e.target.clientHeight + 50;
- }}
- >
- <div
- class="{$settings?.fullScreenMode ?? null
- ? 'max-w-full'
- : 'max-w-2xl md:px-0'} mx-auto w-full px-4"
- >
- <ModelSelector
- bind:selectedModels
- disabled={messages.length > 0 && !selectedModels.includes('')}
- />
- </div>
- <div class=" h-full w-full flex flex-col py-8">
- <Messages
- chatId={$chatId}
- {selectedModels}
- {selectedModelfiles}
- {processing}
- bind:history
- bind:messages
- bind:autoScroll
- bottomPadding={files.length > 0}
- {sendPrompt}
- {continueGeneration}
- {regenerateResponse}
- />
- </div>
- </div>
- <MessageInput
- bind:files
- bind:prompt
- bind:autoScroll
- suggestionPrompts={selectedModelfile?.suggestionPrompts ?? $config.default_prompt_suggestions}
- {messages}
- {submitPrompt}
- {stopResponse}
- />
- </div>
- </div>
|