|
<script lang="ts"> |
|
import { v4 as uuidv4 } from 'uuid'; |
|
import { toast } from 'svelte-sonner'; |
|
import mermaid from 'mermaid'; |
|
|
|
import { getContext, onMount, tick } from 'svelte'; |
|
import { goto } from '$app/navigation'; |
|
import { page } from '$app/stores'; |
|
|
|
import type { Writable } from 'svelte/store'; |
|
import type { i18n as i18nType } from 'i18next'; |
|
import { OLLAMA_API_BASE_URL, OPENAI_API_BASE_URL, WEBUI_BASE_URL } from '$lib/constants'; |
|
|
|
import { |
|
chatId, |
|
chats, |
|
config, |
|
type Model, |
|
models, |
|
settings, |
|
showSidebar, |
|
tags as _tags, |
|
WEBUI_NAME, |
|
banners, |
|
user, |
|
socket, |
|
showCallOverlay, |
|
tools |
|
} from '$lib/stores'; |
|
import { |
|
convertMessagesToHistory, |
|
copyToClipboard, |
|
extractSentencesForAudio, |
|
getUserPosition, |
|
promptTemplate, |
|
splitStream |
|
} from '$lib/utils'; |
|
|
|
import { generateChatCompletion } from '$lib/apis/ollama'; |
|
import { |
|
addTagById, |
|
createNewChat, |
|
deleteTagById, |
|
getAllChatTags, |
|
getChatById, |
|
getChatList, |
|
getTagsById, |
|
updateChatById |
|
} from '$lib/apis/chats'; |
|
import { generateOpenAIChatCompletion } from '$lib/apis/openai'; |
|
import { runWebSearch } from '$lib/apis/rag'; |
|
import { createOpenAITextStream } from '$lib/apis/streaming'; |
|
import { queryMemory } from '$lib/apis/memories'; |
|
import { getAndUpdateUserLocation, getUserSettings } from '$lib/apis/users'; |
|
import { chatCompleted, generateTitle, generateSearchQuery, chatAction } from '$lib/apis'; |
|
|
|
import Banner from '../common/Banner.svelte'; |
|
import MessageInput from '$lib/components/chat/MessageInput.svelte'; |
|
import Messages from '$lib/components/chat/Messages.svelte'; |
|
import Navbar from '$lib/components/layout/Navbar.svelte'; |
|
import CallOverlay from './MessageInput/CallOverlay.svelte'; |
|
import { error } from '@sveltejs/kit'; |
|
import ChatControls from './ChatControls.svelte'; |
|
import EventConfirmDialog from '../common/ConfirmDialog.svelte'; |
|
|
|
const i18n: Writable<i18nType> = getContext('i18n'); |
|
|
|
export let chatIdProp = ''; |
|
let loaded = false; |
|
const eventTarget = new EventTarget(); |
|
|
|
let showControls = false; |
|
let stopResponseFlag = false; |
|
let autoScroll = true; |
|
let processing = ''; |
|
let messagesContainerElement: HTMLDivElement; |
|
|
|
let showEventConfirmation = false; |
|
let eventConfirmationTitle = ''; |
|
let eventConfirmationMessage = ''; |
|
let eventConfirmationInput = false; |
|
let eventConfirmationInputPlaceholder = ''; |
|
let eventConfirmationInputValue = ''; |
|
let eventCallback = null; |
|
|
|
let showModelSelector = true; |
|
|
|
let selectedModels = ['']; |
|
let atSelectedModel: Model | undefined; |
|
|
|
let selectedModelIds = []; |
|
$: selectedModelIds = atSelectedModel !== undefined ? [atSelectedModel.id] : selectedModels; |
|
|
|
let selectedToolIds = []; |
|
let webSearchEnabled = false; |
|
|
|
let chat = null; |
|
let tags = []; |
|
|
|
let title = ''; |
|
let prompt = ''; |
|
|
|
let chatFiles = []; |
|
let files = []; |
|
let messages = []; |
|
let history = { |
|
messages: {}, |
|
currentId: null |
|
}; |
|
|
|
let params = {}; |
|
let valves = {}; |
|
|
|
$: if (history.currentId !== null) { |
|
let _messages = []; |
|
|
|
let currentMessage = history.messages[history.currentId]; |
|
while (currentMessage !== null) { |
|
_messages.unshift({ ...currentMessage }); |
|
currentMessage = |
|
currentMessage.parentId !== null ? history.messages[currentMessage.parentId] : null; |
|
} |
|
messages = _messages; |
|
} else { |
|
messages = []; |
|
} |
|
|
|
$: if (chatIdProp) { |
|
(async () => { |
|
console.log(chatIdProp); |
|
if (chatIdProp && (await loadChat())) { |
|
await tick(); |
|
loaded = true; |
|
|
|
window.setTimeout(() => scrollToBottom(), 0); |
|
const chatInput = document.getElementById('chat-textarea'); |
|
chatInput?.focus(); |
|
} else { |
|
await goto('/'); |
|
} |
|
})(); |
|
} |
|
|
|
const chatEventHandler = async (event, cb) => { |
|
if (event.chat_id === $chatId) { |
|
await tick(); |
|
console.log(event); |
|
let message = history.messages[event.message_id]; |
|
|
|
const type = event?.data?.type ?? null; |
|
const data = event?.data?.data ?? null; |
|
|
|
if (type === 'status') { |
|
if (message?.statusHistory) { |
|
message.statusHistory.push(data); |
|
} else { |
|
message.statusHistory = [data]; |
|
} |
|
} else if (type === 'citation') { |
|
if (message?.citations) { |
|
message.citations.push(data); |
|
} else { |
|
message.citations = [data]; |
|
} |
|
} else if (type === 'message') { |
|
message.content += data.content; |
|
} else if (type === 'replace') { |
|
message.content = data.content; |
|
} else if (type === 'confirmation') { |
|
eventCallback = cb; |
|
|
|
eventConfirmationInput = false; |
|
showEventConfirmation = true; |
|
|
|
eventConfirmationTitle = data.title; |
|
eventConfirmationMessage = data.message; |
|
} else if (type === 'input') { |
|
eventCallback = cb; |
|
|
|
eventConfirmationInput = true; |
|
showEventConfirmation = true; |
|
|
|
eventConfirmationTitle = data.title; |
|
eventConfirmationMessage = data.message; |
|
eventConfirmationInputPlaceholder = data.placeholder; |
|
eventConfirmationInputValue = data?.value ?? ''; |
|
} else { |
|
console.log('Unknown message type', data); |
|
} |
|
|
|
messages = messages; |
|
} |
|
}; |
|
|
|
onMount(async () => { |
|
const onMessageHandler = async (event) => { |
|
if (event.origin === window.origin) { |
|
// Replace with your iframe's origin |
|
console.log('Message received from iframe:', event.data); |
|
if (event.data.type === 'input:prompt') { |
|
console.log(event.data.text); |
|
|
|
const inputElement = document.getElementById('chat-textarea'); |
|
|
|
if (inputElement) { |
|
prompt = event.data.text; |
|
inputElement.focus(); |
|
} |
|
} |
|
|
|
if (event.data.type === 'action:submit') { |
|
console.log(event.data.text); |
|
|
|
if (prompt !== '') { |
|
await tick(); |
|
submitPrompt(prompt); |
|
} |
|
} |
|
|
|
if (event.data.type === 'input:prompt:submit') { |
|
console.log(event.data.text); |
|
|
|
if (prompt !== '') { |
|
await tick(); |
|
submitPrompt(event.data.text); |
|
} |
|
} |
|
} |
|
}; |
|
window.addEventListener('message', onMessageHandler); |
|
|
|
$socket.on('chat-events', chatEventHandler); |
|
|
|
if (!$chatId) { |
|
chatId.subscribe(async (value) => { |
|
if (!value) { |
|
await initNewChat(); |
|
} |
|
}); |
|
} else { |
|
if (!($settings.saveChatHistory ?? true)) { |
|
await goto('/'); |
|
} |
|
} |
|
|
|
return () => { |
|
window.removeEventListener('message', onMessageHandler); |
|
|
|
$socket.off('chat-events'); |
|
}; |
|
}); |
|
|
|
|
|
|
|
|
|
|
|
const initNewChat = async () => { |
|
window.history.replaceState(history.state, '', `/`); |
|
await chatId.set(''); |
|
|
|
autoScroll = true; |
|
|
|
title = ''; |
|
messages = []; |
|
history = { |
|
messages: {}, |
|
currentId: null |
|
}; |
|
|
|
chatFiles = []; |
|
params = {}; |
|
|
|
if ($page.url.searchParams.get('models')) { |
|
selectedModels = $page.url.searchParams.get('models')?.split(','); |
|
} else if ($settings?.models) { |
|
selectedModels = $settings?.models; |
|
} else if ($config?.default_models) { |
|
console.log($config?.default_models.split(',') ?? ''); |
|
selectedModels = $config?.default_models.split(','); |
|
} else { |
|
selectedModels = ['']; |
|
} |
|
|
|
if ($page.url.searchParams.get('q')) { |
|
prompt = $page.url.searchParams.get('q') ?? ''; |
|
|
|
if (prompt) { |
|
await tick(); |
|
submitPrompt(prompt); |
|
} |
|
} |
|
|
|
selectedModels = selectedModels.map((modelId) => |
|
$models.map((m) => m.id).includes(modelId) ? modelId : '' |
|
); |
|
|
|
const userSettings = await getUserSettings(localStorage.token); |
|
|
|
if (userSettings) { |
|
settings.set(userSettings.ui); |
|
} else { |
|
settings.set(JSON.parse(localStorage.getItem('settings') ?? '{}')); |
|
} |
|
|
|
const chatInput = document.getElementById('chat-textarea'); |
|
setTimeout(() => chatInput?.focus(), 0); |
|
}; |
|
|
|
const loadChat = async () => { |
|
chatId.set(chatIdProp); |
|
chat = await getChatById(localStorage.token, $chatId).catch(async (error) => { |
|
await goto('/'); |
|
return null; |
|
}); |
|
|
|
if (chat) { |
|
tags = await getTags(); |
|
const chatContent = chat.chat; |
|
|
|
if (chatContent) { |
|
console.log(chatContent); |
|
|
|
selectedModels = |
|
(chatContent?.models ?? undefined) !== undefined |
|
? chatContent.models |
|
: [chatContent.models ?? '']; |
|
history = |
|
(chatContent?.history ?? undefined) !== undefined |
|
? chatContent.history |
|
: convertMessagesToHistory(chatContent.messages); |
|
title = chatContent.title; |
|
|
|
const userSettings = await getUserSettings(localStorage.token); |
|
|
|
if (userSettings) { |
|
await settings.set(userSettings.ui); |
|
} else { |
|
await settings.set(JSON.parse(localStorage.getItem('settings') ?? '{}')); |
|
} |
|
|
|
params = chatContent?.params ?? {}; |
|
chatFiles = chatContent?.files ?? []; |
|
|
|
autoScroll = true; |
|
await tick(); |
|
|
|
if (messages.length > 0) { |
|
history.messages[messages.at(-1).id].done = true; |
|
} |
|
await tick(); |
|
|
|
return true; |
|
} else { |
|
return null; |
|
} |
|
} |
|
}; |
|
|
|
const scrollToBottom = async () => { |
|
await tick(); |
|
if (messagesContainerElement) { |
|
messagesContainerElement.scrollTop = messagesContainerElement.scrollHeight; |
|
} |
|
}; |
|
|
|
const createMessagesList = (responseMessageId) => { |
|
const message = history.messages[responseMessageId]; |
|
if (message.parentId) { |
|
return [...createMessagesList(message.parentId), message]; |
|
} else { |
|
return [message]; |
|
} |
|
}; |
|
|
|
const chatCompletedHandler = async (chatId, modelId, responseMessageId, messages) => { |
|
await mermaid.run({ |
|
querySelector: '.mermaid' |
|
}); |
|
|
|
const res = await chatCompleted(localStorage.token, { |
|
model: modelId, |
|
messages: messages.map((m) => ({ |
|
id: m.id, |
|
role: m.role, |
|
content: m.content, |
|
info: m.info ? m.info : undefined, |
|
timestamp: m.timestamp |
|
})), |
|
chat_id: chatId, |
|
session_id: $socket?.id, |
|
id: responseMessageId |
|
}).catch((error) => { |
|
toast.error(error); |
|
messages.at(-1).error = { content: error }; |
|
|
|
return null; |
|
}); |
|
|
|
if (res !== null) { |
|
// Update chat history with the new messages |
|
for (const message of res.messages) { |
|
history.messages[message.id] = { |
|
...history.messages[message.id], |
|
...(history.messages[message.id].content !== message.content |
|
? { originalContent: history.messages[message.id].content } |
|
: {}), |
|
...message |
|
}; |
|
} |
|
} |
|
|
|
if ($chatId == chatId) { |
|
if ($settings.saveChatHistory ?? true) { |
|
chat = await updateChatById(localStorage.token, chatId, { |
|
models: selectedModels, |
|
messages: messages, |
|
history: history, |
|
params: params, |
|
files: chatFiles |
|
}); |
|
await chats.set(await getChatList(localStorage.token)); |
|
} |
|
} |
|
}; |
|
|
|
const chatActionHandler = async (chatId, actionId, modelId, responseMessageId) => { |
|
const res = await chatAction(localStorage.token, actionId, { |
|
model: modelId, |
|
messages: messages.map((m) => ({ |
|
id: m.id, |
|
role: m.role, |
|
content: m.content, |
|
info: m.info ? m.info : undefined, |
|
timestamp: m.timestamp |
|
})), |
|
chat_id: chatId, |
|
session_id: $socket?.id, |
|
id: responseMessageId |
|
}).catch((error) => { |
|
toast.error(error); |
|
messages.at(-1).error = { content: error }; |
|
return null; |
|
}); |
|
|
|
if (res !== null) { |
|
// Update chat history with the new messages |
|
for (const message of res.messages) { |
|
history.messages[message.id] = { |
|
...history.messages[message.id], |
|
...(history.messages[message.id].content !== message.content |
|
? { originalContent: history.messages[message.id].content } |
|
: {}), |
|
...message |
|
}; |
|
} |
|
} |
|
|
|
if ($chatId == chatId) { |
|
if ($settings.saveChatHistory ?? true) { |
|
chat = await updateChatById(localStorage.token, chatId, { |
|
models: selectedModels, |
|
messages: messages, |
|
history: history, |
|
params: params, |
|
files: chatFiles |
|
}); |
|
await chats.set(await getChatList(localStorage.token)); |
|
} |
|
} |
|
}; |
|
|
|
const getChatEventEmitter = async (modelId: string, chatId: string = '') => { |
|
return setInterval(() => { |
|
$socket?.emit('usage', { |
|
action: 'chat', |
|
model: modelId, |
|
chat_id: chatId |
|
}); |
|
}, 1000); |
|
}; |
|
|
|
|
|
|
|
|
|
|
|
const submitPrompt = async (userPrompt, { _raw = false } = {}) => { |
|
let _responses = []; |
|
console.log('submitPrompt', $chatId); |
|
|
|
selectedModels = selectedModels.map((modelId) => |
|
$models.map((m) => m.id).includes(modelId) ? modelId : '' |
|
); |
|
|
|
if (selectedModels.includes('')) { |
|
toast.error($i18n.t('Model not selected')); |
|
} else if (messages.length != 0 && messages.at(-1).done != true) { |
|
// Response not done |
|
console.log('wait'); |
|
} else if (messages.length != 0 && messages.at(-1).error) { |
|
// Error in response |
|
toast.error( |
|
$i18n.t( |
|
`Oops! There was an error in the previous response. Please try again or contact admin.` |
|
) |
|
); |
|
} else if ( |
|
files.length > 0 && |
|
files.filter((file) => file.type !== 'image' && file.status !== 'processed').length > 0 |
|
) { |
|
// Upload not done |
|
toast.error( |
|
$i18n.t( |
|
`Oops! Hold tight! Your files are still in the processing oven. We're cooking them up to perfection. Please be patient and we'll let you know once they're ready.` |
|
) |
|
); |
|
} else { |
|
// Reset chat input textarea |
|
const chatTextAreaElement = document.getElementById('chat-textarea'); |
|
|
|
if (chatTextAreaElement) { |
|
chatTextAreaElement.value = ''; |
|
chatTextAreaElement.style.height = ''; |
|
} |
|
|
|
const _files = JSON.parse(JSON.stringify(files)); |
|
chatFiles.push(..._files.filter((item) => ['doc', 'file', 'collection'].includes(item.type))); |
|
chatFiles = chatFiles.filter( |
|
|
|
(item, index, array) => |
|
array.findIndex((i) => JSON.stringify(i) === JSON.stringify(item)) === index |
|
); |
|
|
|
files = []; |
|
|
|
prompt = ''; |
|
|
|
|
|
let userMessageId = uuidv4(); |
|
let userMessage = { |
|
id: userMessageId, |
|
parentId: messages.length !== 0 ? messages.at(-1).id : null, |
|
childrenIds: [], |
|
role: 'user', |
|
content: userPrompt, |
|
files: _files.length > 0 ? _files : undefined, |
|
timestamp: Math.floor(Date.now() / 1000), // Unix epoch |
|
models: selectedModels.filter((m, mIdx) => selectedModels.indexOf(m) === mIdx) |
|
}; |
|
|
|
|
|
history.messages[userMessageId] = userMessage; |
|
history.currentId = userMessageId; |
|
|
|
|
|
if (messages.length !== 0) { |
|
history.messages[messages.at(-1).id].childrenIds.push(userMessageId); |
|
} |
|
|
|
|
|
await tick(); |
|
_responses = await sendPrompt(userPrompt, userMessageId, { newChat: true }); |
|
} |
|
|
|
return _responses; |
|
}; |
|
|
|
const sendPrompt = async (prompt, parentId, { modelId = null, newChat = false } = {}) => { |
|
let _responses = []; |
|
|
|
// If modelId is provided, use it, else use selected model |
|
let selectedModelIds = modelId |
|
? [modelId] |
|
: atSelectedModel !== undefined |
|
? [atSelectedModel.id] |
|
: selectedModels; |
|
|
|
// Create response messages for each selected model |
|
const responseMessageIds = {}; |
|
for (const modelId of selectedModelIds) { |
|
const model = $models.filter((m) => m.id === modelId).at(0); |
|
|
|
if (model) { |
|
let responseMessageId = uuidv4(); |
|
let responseMessage = { |
|
parentId: parentId, |
|
id: responseMessageId, |
|
childrenIds: [], |
|
role: 'assistant', |
|
content: '', |
|
model: model.id, |
|
modelName: model.name ?? model.id, |
|
userContext: null, |
|
timestamp: Math.floor(Date.now() / 1000) // Unix epoch |
|
}; |
|
|
|
|
|
history.messages[responseMessageId] = responseMessage; |
|
history.currentId = responseMessageId; |
|
|
|
|
|
if (parentId !== null) { |
|
history.messages[parentId].childrenIds = [ |
|
...history.messages[parentId].childrenIds, |
|
responseMessageId |
|
]; |
|
} |
|
|
|
responseMessageIds[modelId] = responseMessageId; |
|
} |
|
} |
|
await tick(); |
|
|
|
|
|
if (newChat && messages.length == 2) { |
|
if ($settings.saveChatHistory ?? true) { |
|
chat = await createNewChat(localStorage.token, { |
|
id: $chatId, |
|
title: $i18n.t('New Chat'), |
|
models: selectedModels, |
|
system: $settings.system ?? undefined, |
|
params: params, |
|
messages: messages, |
|
history: history, |
|
tags: [], |
|
timestamp: Date.now() |
|
}); |
|
await chats.set(await getChatList(localStorage.token)); |
|
await chatId.set(chat.id); |
|
} else { |
|
await chatId.set('local'); |
|
} |
|
await tick(); |
|
} |
|
|
|
const _chatId = JSON.parse(JSON.stringify($chatId)); |
|
|
|
await Promise.all( |
|
selectedModelIds.map(async (modelId) => { |
|
console.log('modelId', modelId); |
|
const model = $models.filter((m) => m.id === modelId).at(0); |
|
|
|
if (model) { |
|
// If there are image files, check if model is vision capable |
|
const hasImages = messages.some((message) => |
|
message.files?.some((file) => file.type === 'image') |
|
); |
|
|
|
if (hasImages && !(model.info?.meta?.capabilities?.vision ?? true)) { |
|
toast.error( |
|
$i18n.t('Model {{modelName}} is not vision capable', { |
|
modelName: model.name ?? model.id |
|
}) |
|
); |
|
} |
|
|
|
let responseMessageId = responseMessageIds[modelId]; |
|
let responseMessage = history.messages[responseMessageId]; |
|
|
|
let userContext = null; |
|
if ($settings?.memory ?? false) { |
|
if (userContext === null) { |
|
const res = await queryMemory(localStorage.token, prompt).catch((error) => { |
|
toast.error(error); |
|
return null; |
|
}); |
|
if (res) { |
|
if (res.documents[0].length > 0) { |
|
userContext = res.documents[0].reduce((acc, doc, index) => { |
|
const createdAtTimestamp = res.metadatas[0][index].created_at; |
|
const createdAtDate = new Date(createdAtTimestamp * 1000) |
|
.toISOString() |
|
.split('T')[0]; |
|
return `${acc}${index + 1}. [${createdAtDate}]. ${doc}\n`; |
|
}, ''); |
|
} |
|
|
|
console.log(userContext); |
|
} |
|
} |
|
} |
|
responseMessage.userContext = userContext; |
|
|
|
const chatEventEmitter = await getChatEventEmitter(model.id, _chatId); |
|
if (webSearchEnabled) { |
|
await getWebSearchResults(model.id, parentId, responseMessageId); |
|
} |
|
|
|
let _response = null; |
|
if (model?.owned_by === 'openai') { |
|
_response = await sendPromptOpenAI(model, prompt, responseMessageId, _chatId); |
|
} else if (model) { |
|
_response = await sendPromptOllama(model, prompt, responseMessageId, _chatId); |
|
} |
|
_responses.push(_response); |
|
|
|
if (chatEventEmitter) clearInterval(chatEventEmitter); |
|
} else { |
|
toast.error($i18n.t(`Model {{modelId}} not found`, { modelId })); |
|
} |
|
}) |
|
); |
|
|
|
await chats.set(await getChatList(localStorage.token)); |
|
return _responses; |
|
}; |
|
|
|
const sendPromptOllama = async (model, userPrompt, responseMessageId, _chatId) => { |
|
let _response = null; |
|
|
|
const responseMessage = history.messages[responseMessageId]; |
|
const userMessage = history.messages[responseMessage.parentId]; |
|
|
|
// Wait until history/message have been updated |
|
await tick(); |
|
|
|
// Scroll down |
|
scrollToBottom(); |
|
|
|
const messagesBody = [ |
|
params?.system || $settings.system || (responseMessage?.userContext ?? null) |
|
? { |
|
role: 'system', |
|
content: `${promptTemplate( |
|
params?.system ?? $settings?.system ?? '', |
|
$user.name, |
|
$settings?.userLocation |
|
? await getAndUpdateUserLocation(localStorage.token) |
|
: undefined |
|
)}${ |
|
responseMessage?.userContext ?? null |
|
? `\n\nUser Context:\n${responseMessage?.userContext ?? ''}` |
|
: '' |
|
}` |
|
} |
|
: undefined, |
|
...messages |
|
] |
|
.filter((message) => message?.content?.trim()) |
|
.map((message, idx, arr) => { |
|
// Prepare the base message object |
|
const baseMessage = { |
|
role: message.role, |
|
content: message.content |
|
}; |
|
|
|
|
|
const imageUrls = message.files |
|
?.filter((file) => file.type === 'image') |
|
.map((file) => file.url.slice(file.url.indexOf(',') + 1)); |
|
|
|
|
|
if (imageUrls && imageUrls.length > 0 && message.role === 'user') { |
|
baseMessage.images = imageUrls; |
|
} |
|
return baseMessage; |
|
}); |
|
|
|
let lastImageIndex = -1; |
|
|
|
|
|
messagesBody.forEach((item, index) => { |
|
if (item.images) { |
|
lastImageIndex = index; |
|
} |
|
}); |
|
|
|
|
|
messagesBody.forEach((item, index) => { |
|
if (index !== lastImageIndex) { |
|
delete item.images; |
|
} |
|
}); |
|
|
|
let files = JSON.parse(JSON.stringify(chatFiles)); |
|
if (model?.info?.meta?.knowledge ?? false) { |
|
files.push(...model.info.meta.knowledge); |
|
} |
|
files.push( |
|
...(userMessage?.files ?? []).filter((item) => |
|
['doc', 'file', 'collection'].includes(item.type) |
|
), |
|
...(responseMessage?.files ?? []).filter((item) => ['web_search_results'].includes(item.type)) |
|
); |
|
|
|
eventTarget.dispatchEvent( |
|
new CustomEvent('chat:start', { |
|
detail: { |
|
id: responseMessageId |
|
} |
|
}) |
|
); |
|
|
|
await tick(); |
|
|
|
const [res, controller] = await generateChatCompletion(localStorage.token, { |
|
stream: true, |
|
model: model.id, |
|
messages: messagesBody, |
|
options: { |
|
...(params ?? $settings.params ?? {}), |
|
stop: |
|
params?.stop ?? $settings?.params?.stop ?? undefined |
|
? (params?.stop ?? $settings.params.stop).map((str) => |
|
decodeURIComponent(JSON.parse('"' + str.replace(/\"/g, '\\"') + '"')) |
|
) |
|
: undefined, |
|
num_predict: params?.max_tokens ?? $settings?.params?.max_tokens ?? undefined, |
|
repeat_penalty: |
|
params?.frequency_penalty ?? $settings?.params?.frequency_penalty ?? undefined |
|
}, |
|
format: $settings.requestFormat ?? undefined, |
|
keep_alive: $settings.keepAlive ?? undefined, |
|
tool_ids: selectedToolIds.length > 0 ? selectedToolIds : undefined, |
|
files: files.length > 0 ? files : undefined, |
|
...(Object.keys(valves).length ? { valves } : {}), |
|
session_id: $socket?.id, |
|
chat_id: $chatId, |
|
id: responseMessageId |
|
}); |
|
|
|
if (res && res.ok) { |
|
console.log('controller', controller); |
|
|
|
const reader = res.body |
|
.pipeThrough(new TextDecoderStream()) |
|
.pipeThrough(splitStream('\n')) |
|
.getReader(); |
|
|
|
while (true) { |
|
const { value, done } = await reader.read(); |
|
if (done || stopResponseFlag || _chatId !== $chatId) { |
|
responseMessage.done = true; |
|
messages = messages; |
|
|
|
if (stopResponseFlag) { |
|
controller.abort('User: Stop Response'); |
|
} else { |
|
const messages = createMessagesList(responseMessageId); |
|
await chatCompletedHandler(_chatId, model.id, responseMessageId, messages); |
|
} |
|
|
|
_response = responseMessage.content; |
|
break; |
|
} |
|
|
|
try { |
|
let lines = value.split('\n'); |
|
|
|
for (const line of lines) { |
|
if (line !== '') { |
|
console.log(line); |
|
let data = JSON.parse(line); |
|
|
|
if ('citations' in data) { |
|
responseMessage.citations = data.citations; |
|
continue; |
|
} |
|
|
|
if ('detail' in data) { |
|
throw data; |
|
} |
|
|
|
if (data.done == false) { |
|
if (responseMessage.content == '' && data.message.content == '\n') { |
|
continue; |
|
} else { |
|
responseMessage.content += data.message.content; |
|
|
|
const sentences = extractSentencesForAudio(responseMessage.content); |
|
sentences.pop(); |
|
|
|
// dispatch only last sentence and make sure it hasn't been dispatched before |
|
if ( |
|
sentences.length > 0 && |
|
sentences[sentences.length - 1] !== responseMessage.lastSentence |
|
) { |
|
responseMessage.lastSentence = sentences[sentences.length - 1]; |
|
eventTarget.dispatchEvent( |
|
new CustomEvent('chat', { |
|
detail: { id: responseMessageId, content: sentences[sentences.length - 1] } |
|
}) |
|
); |
|
} |
|
|
|
messages = messages; |
|
} |
|
} else { |
|
responseMessage.done = true; |
|
|
|
if (responseMessage.content == '') { |
|
responseMessage.error = { |
|
code: 400, |
|
content: `Oops! No text generated from Ollama, Please try again.` |
|
}; |
|
} |
|
|
|
responseMessage.context = data.context ?? null; |
|
responseMessage.info = { |
|
total_duration: data.total_duration, |
|
load_duration: data.load_duration, |
|
sample_count: data.sample_count, |
|
sample_duration: data.sample_duration, |
|
prompt_eval_count: data.prompt_eval_count, |
|
prompt_eval_duration: data.prompt_eval_duration, |
|
eval_count: data.eval_count, |
|
eval_duration: data.eval_duration |
|
}; |
|
messages = messages; |
|
|
|
if ($settings.notificationEnabled && !document.hasFocus()) { |
|
const notification = new Notification(`${model.id}`, { |
|
body: responseMessage.content, |
|
icon: `${WEBUI_BASE_URL}/static/favicon.png` |
|
}); |
|
} |
|
|
|
if ($settings?.responseAutoCopy ?? false) { |
|
copyToClipboard(responseMessage.content); |
|
} |
|
|
|
if ($settings.responseAutoPlayback && !$showCallOverlay) { |
|
await tick(); |
|
document.getElementById(`speak-button-${responseMessage.id}`)?.click(); |
|
} |
|
} |
|
} |
|
} |
|
} catch (error) { |
|
console.log(error); |
|
if ('detail' in error) { |
|
toast.error(error.detail); |
|
} |
|
break; |
|
} |
|
|
|
if (autoScroll) { |
|
scrollToBottom(); |
|
} |
|
} |
|
|
|
if ($chatId == _chatId) { |
|
if ($settings.saveChatHistory ?? true) { |
|
chat = await updateChatById(localStorage.token, _chatId, { |
|
messages: messages, |
|
history: history, |
|
models: selectedModels, |
|
params: params, |
|
files: chatFiles |
|
}); |
|
await chats.set(await getChatList(localStorage.token)); |
|
} |
|
} |
|
} else { |
|
if (res !== null) { |
|
const error = await res.json(); |
|
console.log(error); |
|
if ('detail' in error) { |
|
toast.error(error.detail); |
|
responseMessage.error = { content: error.detail }; |
|
} else { |
|
toast.error(error.error); |
|
responseMessage.error = { content: error.error }; |
|
} |
|
} else { |
|
toast.error( |
|
$i18n.t(`Uh-oh! There was an issue connecting to {{provider}}.`, { provider: 'Ollama' }) |
|
); |
|
responseMessage.error = { |
|
content: $i18n.t(`Uh-oh! There was an issue connecting to {{provider}}.`, { |
|
provider: 'Ollama' |
|
}) |
|
}; |
|
} |
|
responseMessage.done = true; |
|
messages = messages; |
|
} |
|
|
|
stopResponseFlag = false; |
|
await tick(); |
|
|
|
let lastSentence = extractSentencesForAudio(responseMessage.content)?.at(-1) ?? ''; |
|
if (lastSentence) { |
|
eventTarget.dispatchEvent( |
|
new CustomEvent('chat', { |
|
detail: { id: responseMessageId, content: lastSentence } |
|
}) |
|
); |
|
} |
|
eventTarget.dispatchEvent( |
|
new CustomEvent('chat:finish', { |
|
detail: { |
|
id: responseMessageId, |
|
content: responseMessage.content |
|
} |
|
}) |
|
); |
|
|
|
if (autoScroll) { |
|
scrollToBottom(); |
|
} |
|
|
|
if (messages.length == 2 && messages.at(1).content !== '' && selectedModels[0] === model.id) { |
|
window.history.replaceState(history.state, '', `/c/${_chatId}`); |
|
const _title = await generateChatTitle(userPrompt); |
|
await setChatTitle(_chatId, _title); |
|
} |
|
|
|
return _response; |
|
}; |
|
|
|
const sendPromptOpenAI = async (model, userPrompt, responseMessageId, _chatId) => { |
|
let _response = null; |
|
|
|
const responseMessage = history.messages[responseMessageId]; |
|
const userMessage = history.messages[responseMessage.parentId]; |
|
|
|
let files = JSON.parse(JSON.stringify(chatFiles)); |
|
if (model?.info?.meta?.knowledge ?? false) { |
|
files.push(...model.info.meta.knowledge); |
|
} |
|
files.push( |
|
...(userMessage?.files ?? []).filter((item) => |
|
['doc', 'file', 'collection'].includes(item.type) |
|
), |
|
...(responseMessage?.files ?? []).filter((item) => ['web_search_results'].includes(item.type)) |
|
); |
|
|
|
scrollToBottom(); |
|
|
|
eventTarget.dispatchEvent( |
|
new CustomEvent('chat:start', { |
|
detail: { |
|
id: responseMessageId |
|
} |
|
}) |
|
); |
|
await tick(); |
|
|
|
try { |
|
const [res, controller] = await generateOpenAIChatCompletion( |
|
localStorage.token, |
|
{ |
|
stream: true, |
|
model: model.id, |
|
stream_options: |
|
model.info?.meta?.capabilities?.usage ?? false |
|
? { |
|
include_usage: true |
|
} |
|
: undefined, |
|
messages: [ |
|
params?.system || $settings.system || (responseMessage?.userContext ?? null) |
|
? { |
|
role: 'system', |
|
content: `${promptTemplate( |
|
params?.system ?? $settings?.system ?? '', |
|
$user.name, |
|
$settings?.userLocation |
|
? await getAndUpdateUserLocation(localStorage.token) |
|
: undefined |
|
)}${ |
|
responseMessage?.userContext ?? null |
|
? `\n\nUser Context:\n${responseMessage?.userContext ?? ''}` |
|
: '' |
|
}` |
|
} |
|
: undefined, |
|
...messages |
|
] |
|
.filter((message) => message?.content?.trim()) |
|
.map((message, idx, arr) => ({ |
|
role: message.role, |
|
...((message.files?.filter((file) => file.type === 'image').length > 0 ?? false) && |
|
message.role === 'user' |
|
? { |
|
content: [ |
|
{ |
|
type: 'text', |
|
text: |
|
arr.length - 1 !== idx |
|
? message.content |
|
: message?.raContent ?? message.content |
|
}, |
|
...message.files |
|
.filter((file) => file.type === 'image') |
|
.map((file) => ({ |
|
type: 'image_url', |
|
image_url: { |
|
url: file.url |
|
} |
|
})) |
|
] |
|
} |
|
: { |
|
content: |
|
arr.length - 1 !== idx |
|
? message.content |
|
: message?.raContent ?? message.content |
|
}) |
|
})), |
|
seed: params?.seed ?? $settings?.params?.seed ?? undefined, |
|
stop: |
|
params?.stop ?? $settings?.params?.stop ?? undefined |
|
? (params?.stop ?? $settings.params.stop).map((str) => |
|
decodeURIComponent(JSON.parse('"' + str.replace(/\"/g, '\\"') + '"')) |
|
) |
|
: undefined, |
|
temperature: params?.temperature ?? $settings?.params?.temperature ?? undefined, |
|
top_p: params?.top_p ?? $settings?.params?.top_p ?? undefined, |
|
frequency_penalty: |
|
params?.frequency_penalty ?? $settings?.params?.frequency_penalty ?? undefined, |
|
max_tokens: params?.max_tokens ?? $settings?.params?.max_tokens ?? undefined, |
|
tool_ids: selectedToolIds.length > 0 ? selectedToolIds : undefined, |
|
files: files.length > 0 ? files : undefined, |
|
...(Object.keys(valves).length ? { valves } : {}), |
|
session_id: $socket?.id, |
|
chat_id: $chatId, |
|
id: responseMessageId |
|
}, |
|
`${WEBUI_BASE_URL}/api` |
|
); |
|
|
|
// Wait until history/message have been updated |
|
await tick(); |
|
|
|
scrollToBottom(); |
|
|
|
if (res && res.ok && res.body) { |
|
const textStream = await createOpenAITextStream(res.body, $settings.splitLargeChunks); |
|
let lastUsage = null; |
|
|
|
for await (const update of textStream) { |
|
const { value, done, citations, error, usage } = update; |
|
if (error) { |
|
await handleOpenAIError(error, null, model, responseMessage); |
|
break; |
|
} |
|
if (done || stopResponseFlag || _chatId !== $chatId) { |
|
responseMessage.done = true; |
|
messages = messages; |
|
|
|
if (stopResponseFlag) { |
|
controller.abort('User: Stop Response'); |
|
} else { |
|
const messages = createMessagesList(responseMessageId); |
|
|
|
await chatCompletedHandler(_chatId, model.id, responseMessageId, messages); |
|
} |
|
|
|
_response = responseMessage.content; |
|
|
|
break; |
|
} |
|
|
|
if (usage) { |
|
lastUsage = usage; |
|
} |
|
|
|
if (citations) { |
|
responseMessage.citations = citations; |
|
continue; |
|
} |
|
|
|
if (responseMessage.content == '' && value == '\n') { |
|
continue; |
|
} else { |
|
responseMessage.content += value; |
|
|
|
const sentences = extractSentencesForAudio(responseMessage.content); |
|
sentences.pop(); |
|
|
|
// dispatch only last sentence and make sure it hasn't been dispatched before |
|
if ( |
|
sentences.length > 0 && |
|
sentences[sentences.length - 1] !== responseMessage.lastSentence |
|
) { |
|
responseMessage.lastSentence = sentences[sentences.length - 1]; |
|
eventTarget.dispatchEvent( |
|
new CustomEvent('chat', { |
|
detail: { id: responseMessageId, content: sentences[sentences.length - 1] } |
|
}) |
|
); |
|
} |
|
|
|
messages = messages; |
|
} |
|
|
|
if (autoScroll) { |
|
scrollToBottom(); |
|
} |
|
} |
|
|
|
if ($settings.notificationEnabled && !document.hasFocus()) { |
|
const notification = new Notification(`${model.id}`, { |
|
body: responseMessage.content, |
|
icon: `${WEBUI_BASE_URL}/static/favicon.png` |
|
}); |
|
} |
|
|
|
if ($settings.responseAutoCopy) { |
|
copyToClipboard(responseMessage.content); |
|
} |
|
|
|
if ($settings.responseAutoPlayback && !$showCallOverlay) { |
|
await tick(); |
|
|
|
document.getElementById(`speak-button-${responseMessage.id}`)?.click(); |
|
} |
|
|
|
if (lastUsage) { |
|
responseMessage.info = { ...lastUsage, openai: true }; |
|
} |
|
|
|
if ($chatId == _chatId) { |
|
if ($settings.saveChatHistory ?? true) { |
|
chat = await updateChatById(localStorage.token, _chatId, { |
|
models: selectedModels, |
|
messages: messages, |
|
history: history, |
|
params: params, |
|
files: chatFiles |
|
}); |
|
await chats.set(await getChatList(localStorage.token)); |
|
} |
|
} |
|
} else { |
|
await handleOpenAIError(null, res, model, responseMessage); |
|
} |
|
} catch (error) { |
|
await handleOpenAIError(error, null, model, responseMessage); |
|
} |
|
messages = messages; |
|
|
|
stopResponseFlag = false; |
|
await tick(); |
|
|
|
let lastSentence = extractSentencesForAudio(responseMessage.content)?.at(-1) ?? ''; |
|
if (lastSentence) { |
|
eventTarget.dispatchEvent( |
|
new CustomEvent('chat', { |
|
detail: { id: responseMessageId, content: lastSentence } |
|
}) |
|
); |
|
} |
|
|
|
eventTarget.dispatchEvent( |
|
new CustomEvent('chat:finish', { |
|
detail: { |
|
id: responseMessageId, |
|
content: responseMessage.content |
|
} |
|
}) |
|
); |
|
|
|
if (autoScroll) { |
|
scrollToBottom(); |
|
} |
|
|
|
if (messages.length == 2 && selectedModels[0] === model.id) { |
|
window.history.replaceState(history.state, '', `/c/${_chatId}`); |
|
|
|
const _title = await generateChatTitle(userPrompt); |
|
await setChatTitle(_chatId, _title); |
|
} |
|
|
|
return _response; |
|
}; |
|
|
|
const handleOpenAIError = async (error, res: Response | null, model, responseMessage) => { |
|
let errorMessage = ''; |
|
let innerError; |
|
|
|
if (error) { |
|
innerError = error; |
|
} else if (res !== null) { |
|
innerError = await res.json(); |
|
} |
|
console.error(innerError); |
|
if ('detail' in innerError) { |
|
toast.error(innerError.detail); |
|
errorMessage = innerError.detail; |
|
} else if ('error' in innerError) { |
|
if ('message' in innerError.error) { |
|
toast.error(innerError.error.message); |
|
errorMessage = innerError.error.message; |
|
} else { |
|
toast.error(innerError.error); |
|
errorMessage = innerError.error; |
|
} |
|
} else if ('message' in innerError) { |
|
toast.error(innerError.message); |
|
errorMessage = innerError.message; |
|
} |
|
|
|
responseMessage.error = { |
|
content: |
|
$i18n.t(`Uh-oh! There was an issue connecting to {{provider}}.`, { |
|
provider: model.name ?? model.id |
|
}) + |
|
'\n' + |
|
errorMessage |
|
}; |
|
responseMessage.done = true; |
|
|
|
messages = messages; |
|
}; |
|
|
|
const stopResponse = () => { |
|
stopResponseFlag = true; |
|
console.log('stopResponse'); |
|
}; |
|
|
|
const regenerateResponse = async (message) => { |
|
console.log('regenerateResponse'); |
|
|
|
if (messages.length != 0) { |
|
let userMessage = history.messages[message.parentId]; |
|
let userPrompt = userMessage.content; |
|
|
|
if ((userMessage?.models ?? [...selectedModels]).length == 1) { |
|
// If user message has only one model selected, sendPrompt automatically selects it for regeneration |
|
await sendPrompt(userPrompt, userMessage.id); |
|
} else { |
|
// If there are multiple models selected, use the model of the response message for regeneration |
|
// e.g. many model chat |
|
await sendPrompt(userPrompt, userMessage.id, { modelId: message.model }); |
|
} |
|
} |
|
}; |
|
|
|
const continueGeneration = async () => { |
|
console.log('continueGeneration'); |
|
const _chatId = JSON.parse(JSON.stringify($chatId)); |
|
|
|
if (messages.length != 0 && messages.at(-1).done == true) { |
|
const responseMessage = history.messages[history.currentId]; |
|
responseMessage.done = false; |
|
await tick(); |
|
|
|
const model = $models.filter((m) => m.id === responseMessage.model).at(0); |
|
|
|
if (model) { |
|
if (model?.owned_by === 'openai') { |
|
await sendPromptOpenAI( |
|
model, |
|
history.messages[responseMessage.parentId].content, |
|
responseMessage.id, |
|
_chatId |
|
); |
|
} else |
|
await sendPromptOllama( |
|
model, |
|
history.messages[responseMessage.parentId].content, |
|
responseMessage.id, |
|
_chatId |
|
); |
|
} |
|
} else { |
|
toast.error($i18n.t(`Model {{modelId}} not found`, { modelId })); |
|
} |
|
}; |
|
|
|
const generateChatTitle = async (userPrompt) => { |
|
if ($settings?.title?.auto ?? true) { |
|
const title = await generateTitle( |
|
localStorage.token, |
|
selectedModels[0], |
|
userPrompt, |
|
$chatId |
|
).catch((error) => { |
|
console.error(error); |
|
return 'New Chat'; |
|
}); |
|
|
|
return title; |
|
} else { |
|
return `${userPrompt}`; |
|
} |
|
}; |
|
|
|
const setChatTitle = async (_chatId, _title) => { |
|
if (_chatId === $chatId) { |
|
title = _title; |
|
} |
|
|
|
if ($settings.saveChatHistory ?? true) { |
|
chat = await updateChatById(localStorage.token, _chatId, { title: _title }); |
|
await chats.set(await getChatList(localStorage.token)); |
|
} |
|
}; |
|
|
|
const getWebSearchResults = async (model: string, parentId: string, responseId: string) => { |
|
const responseMessage = history.messages[responseId]; |
|
const userMessage = history.messages[parentId]; |
|
|
|
responseMessage.statusHistory = [ |
|
{ |
|
done: false, |
|
action: 'web_search', |
|
description: $i18n.t('Generating search query') |
|
} |
|
]; |
|
messages = messages; |
|
|
|
const prompt = userMessage.content; |
|
let searchQuery = await generateSearchQuery(localStorage.token, model, messages, prompt).catch( |
|
(error) => { |
|
console.log(error); |
|
return prompt; |
|
} |
|
); |
|
|
|
if (!searchQuery) { |
|
toast.warning($i18n.t('No search query generated')); |
|
responseMessage.statusHistory.push({ |
|
done: true, |
|
error: true, |
|
action: 'web_search', |
|
description: 'No search query generated' |
|
}); |
|
|
|
messages = messages; |
|
} |
|
|
|
responseMessage.statusHistory.push({ |
|
done: false, |
|
action: 'web_search', |
|
description: $i18n.t(`Searching "{{searchQuery}}"`, { searchQuery }) |
|
}); |
|
messages = messages; |
|
|
|
const results = await runWebSearch(localStorage.token, searchQuery).catch((error) => { |
|
console.log(error); |
|
toast.error(error); |
|
|
|
return null; |
|
}); |
|
|
|
if (results) { |
|
responseMessage.statusHistory.push({ |
|
done: true, |
|
action: 'web_search', |
|
description: $i18n.t('Searched {{count}} sites', { count: results.filenames.length }), |
|
query: searchQuery, |
|
urls: results.filenames |
|
}); |
|
|
|
if (responseMessage?.files ?? undefined === undefined) { |
|
responseMessage.files = []; |
|
} |
|
|
|
responseMessage.files.push({ |
|
collection_name: results.collection_name, |
|
name: searchQuery, |
|
type: 'web_search_results', |
|
urls: results.filenames |
|
}); |
|
|
|
messages = messages; |
|
} else { |
|
responseMessage.statusHistory.push({ |
|
done: true, |
|
error: true, |
|
action: 'web_search', |
|
description: 'No search results found' |
|
}); |
|
messages = messages; |
|
} |
|
}; |
|
|
|
const getTags = async () => { |
|
return await getTagsById(localStorage.token, $chatId).catch(async (error) => { |
|
return []; |
|
}); |
|
}; |
|
</script> |
|
|
|
<svelte:head> |
|
<title> |
|
{title |
|
? `${title.length > 30 ? `${title.slice(0, 30)}...` : title} | ${$WEBUI_NAME}` |
|
: `${$WEBUI_NAME}`} |
|
</title> |
|
</svelte:head> |
|
|
|
<audio id="audioElement" src="" style="display: none;" /> |
|
|
|
<EventConfirmDialog |
|
bind:show={showEventConfirmation} |
|
title={eventConfirmationTitle} |
|
message={eventConfirmationMessage} |
|
input={eventConfirmationInput} |
|
inputPlaceholder={eventConfirmationInputPlaceholder} |
|
inputValue={eventConfirmationInputValue} |
|
on:confirm={(e) => { |
|
if (e.detail) { |
|
eventCallback(e.detail); |
|
} else { |
|
eventCallback(true); |
|
} |
|
}} |
|
on:cancel={() => { |
|
eventCallback(false); |
|
}} |
|
/> |
|
|
|
{#if $showCallOverlay} |
|
<CallOverlay |
|
{submitPrompt} |
|
{stopResponse} |
|
bind:files |
|
modelId={selectedModelIds?.at(0) ?? null} |
|
chatId={$chatId} |
|
{eventTarget} |
|
/> |
|
{/if} |
|
|
|
{#if !chatIdProp || (loaded && chatIdProp)} |
|
<div |
|
class="h-screen max-h-[100dvh] {$showSidebar |
|
? 'md:max-w-[calc(100%-260px)]' |
|
: ''} w-full max-w-full flex flex-col" |
|
> |
|
{#if $settings?.backgroundImageUrl ?? null} |
|
<div |
|
class="absolute {$showSidebar |
|
? 'md:max-w-[calc(100%-260px)] md:translate-x-[260px]' |
|
: ''} top-0 left-0 w-full h-full bg-cover bg-center bg-no-repeat" |
|
style="background-image: url({$settings.backgroundImageUrl}) " |
|
/> |
|
|
|
<div |
|
class="absolute top-0 left-0 w-full h-full bg-gradient-to-t from-white to-white/85 dark:from-gray-900 dark:to-[#171717]/90 z-0" |
|
/> |
|
{/if} |
|
|
|
<Navbar |
|
{title} |
|
bind:selectedModels |
|
bind:showModelSelector |
|
bind:showControls |
|
shareEnabled={messages.length > 0} |
|
{chat} |
|
{initNewChat} |
|
/> |
|
|
|
{#if $banners.length > 0 && messages.length === 0 && !$chatId && selectedModels.length <= 1} |
|
<div |
|
class="absolute top-[4.25rem] w-full {$showSidebar |
|
? 'md:max-w-[calc(100%-260px)]' |
|
: ''} {showControls ? 'lg:pr-[24rem]' : ''} z-20" |
|
> |
|
<div class=" flex flex-col gap-1 w-full"> |
|
{#each $banners.filter( (b) => (b.dismissible ? !JSON.parse(localStorage.getItem('dismissedBannerIds') ?? '[]').includes(b.id) : true) ) as banner} |
|
<Banner |
|
{banner} |
|
on:dismiss={(e) => { |
|
const bannerId = e.detail; |
|
|
|
localStorage.setItem( |
|
'dismissedBannerIds', |
|
JSON.stringify( |
|
[ |
|
bannerId, |
|
...JSON.parse(localStorage.getItem('dismissedBannerIds') ?? '[]') |
|
].filter((id) => $banners.find((b) => b.id === id)) |
|
) |
|
); |
|
}} |
|
/> |
|
{/each} |
|
</div> |
|
</div> |
|
{/if} |
|
|
|
<div class="flex flex-col flex-auto z-10"> |
|
<div |
|
class=" pb-2.5 flex flex-col justify-between w-full flex-auto overflow-auto h-0 max-w-full z-10 scrollbar-hidden {showControls |
|
? 'lg:pr-[24rem]' |
|
: ''}" |
|
id="messages-container" |
|
bind:this={messagesContainerElement} |
|
on:scroll={(e) => { |
|
autoScroll = |
|
messagesContainerElement.scrollHeight - messagesContainerElement.scrollTop <= |
|
messagesContainerElement.clientHeight + 5; |
|
}} |
|
> |
|
<div class=" h-full w-full flex flex-col {chatIdProp ? 'py-4' : 'pt-2 pb-4'}"> |
|
<Messages |
|
chatId={$chatId} |
|
{selectedModels} |
|
{processing} |
|
bind:history |
|
bind:messages |
|
bind:autoScroll |
|
bind:prompt |
|
bottomPadding={files.length > 0} |
|
{sendPrompt} |
|
{continueGeneration} |
|
{regenerateResponse} |
|
{chatActionHandler} |
|
/> |
|
</div> |
|
</div> |
|
|
|
<div class={showControls ? 'lg:pr-[24rem]' : ''}> |
|
<MessageInput |
|
bind:files |
|
bind:prompt |
|
bind:autoScroll |
|
bind:selectedToolIds |
|
bind:webSearchEnabled |
|
bind:atSelectedModel |
|
availableToolIds={selectedModelIds.reduce((a, e, i, arr) => { |
|
const model = $models.find((m) => m.id === e); |
|
if (model?.info?.meta?.toolIds ?? false) { |
|
return [...new Set([...a, ...model.info.meta.toolIds])]; |
|
} |
|
return a; |
|
}, [])} |
|
transparentBackground={$settings?.backgroundImageUrl ?? false} |
|
{selectedModels} |
|
{messages} |
|
{submitPrompt} |
|
{stopResponse} |
|
/> |
|
</div> |
|
</div> |
|
|
|
<ChatControls |
|
models={selectedModelIds.reduce((a, e, i, arr) => { |
|
const model = $models.find((m) => m.id === e); |
|
if (model) { |
|
return [...a, model]; |
|
} |
|
return a; |
|
}, [])} |
|
bind:show={showControls} |
|
bind:chatFiles |
|
bind:params |
|
bind:valves |
|
/> |
|
</div> |
|
{/if} |
|
|