chore: build rollback

main
ChenZhaoYu 2 years ago
parent 78bcf7f4ce
commit 76cef650b4

@ -25,21 +25,12 @@ router.post('/chat-process', [auth, limiter], async (req, res) => {
try { try {
const { prompt, options = {}, systemMessage } = req.body as RequestProps const { prompt, options = {}, systemMessage } = req.body as RequestProps
let firstChunk = true let firstChunk = true
let chatLength = 0
let newChatLength = 0
await chatReplyProcess({ await chatReplyProcess({
message: prompt, message: prompt,
lastContext: options, lastContext: options,
process: (chat: ChatMessage) => { process: (chat: ChatMessage) => {
if (firstChunk) { res.write(firstChunk ? JSON.stringify(chat) : `\n${JSON.stringify(chat)}`)
res.write(`${JSON.stringify(chat)}t1h1i4s5i1s4a1s9i1l9l8y1s0plit`) firstChunk = false
firstChunk = false
}
else if (chatLength !== chat.text.length) {
newChatLength = chat.text.length
res.write(chat.text.substring(chatLength, newChatLength))
chatLength = newChatLength
}
}, },
systemMessage, systemMessage,
}) })

@ -107,9 +107,7 @@ async function onConversation() {
scrollToBottom() scrollToBottom()
try { try {
const magicSplit = 't1h1i4s5i1s4a1s9i1l9l8y1s0plit' let lastText = ''
let renderText = ''
let firstTime = true
const fetchChatAPIOnce = async () => { const fetchChatAPIOnce = async () => {
await fetchChatAPIProcess<Chat.ConversationResponse>({ await fetchChatAPIProcess<Chat.ConversationResponse>({
prompt: message, prompt: message,
@ -119,49 +117,42 @@ async function onConversation() {
const xhr = event.target const xhr = event.target
const { responseText } = xhr const { responseText } = xhr
// Always process the final line // Always process the final line
const lastIndex = responseText.lastIndexOf('\n', responseText.length - 2)
const splitIndexBegin = responseText.search(magicSplit) let chunk = responseText
if (splitIndexBegin !== -1) { if (lastIndex !== -1)
const splitIndexEnd = splitIndexBegin + magicSplit.length chunk = responseText.substring(lastIndex)
try {
const firstChunk = responseText.substring(0, splitIndexBegin) const data = JSON.parse(chunk)
const deltaText = responseText.substring(splitIndexEnd) updateChat(
try { +uuid,
const data = JSON.parse(firstChunk) dataSources.value.length - 1,
if (firstTime) { {
firstTime = false dateTime: new Date().toLocaleString(),
renderText = data.text ?? '' text: lastText + data.text ?? '',
} inversion: false,
else { error: false,
renderText = deltaText ?? '' loading: false,
} conversationOptions: { conversationId: data.conversationId, parentMessageId: data.id },
updateChat( requestOptions: { prompt: message, options: { ...options } },
+uuid, },
dataSources.value.length - 1, )
{
dateTime: new Date().toLocaleString(), if (openLongReply && data.detail.choices[0].finish_reason === 'length') {
text: renderText, options.parentMessageId = data.id
inversion: false, lastText = data.text
error: false, message = ''
loading: false, return fetchChatAPIOnce()
conversationOptions: { conversationId: data.conversationId, parentMessageId: data.id },
requestOptions: { prompt: message, ...options },
},
)
if (openLongReply && data.detail.choices[0].finish_reason === 'length') {
options.parentMessageId = data.id
message = ''
return fetchChatAPIOnce()
}
}
catch (error) {
//
} }
scrollToBottomIfAtBottom()
}
catch (error) {
//
} }
}, },
}) })
} }
await fetchChatAPIOnce() await fetchChatAPIOnce()
} }
catch (error: any) { catch (error: any) {
@ -246,9 +237,7 @@ async function onRegenerate(index: number) {
) )
try { try {
const magicSplit = 't1h1i4s5i1s4a1s9i1l9l8y1s0plit' let lastText = ''
let renderText = ''
let firstTime = true
const fetchChatAPIOnce = async () => { const fetchChatAPIOnce = async () => {
await fetchChatAPIProcess<Chat.ConversationResponse>({ await fetchChatAPIProcess<Chat.ConversationResponse>({
prompt: message, prompt: message,
@ -258,46 +247,36 @@ async function onRegenerate(index: number) {
const xhr = event.target const xhr = event.target
const { responseText } = xhr const { responseText } = xhr
// Always process the final line // Always process the final line
const lastIndex = responseText.lastIndexOf('\n', responseText.length - 2)
const splitIndexBegin = responseText.search(magicSplit) let chunk = responseText
if (splitIndexBegin !== -1) { if (lastIndex !== -1)
const splitIndexEnd = splitIndexBegin + magicSplit.length chunk = responseText.substring(lastIndex)
try {
const firstChunk = responseText.substring(0, splitIndexBegin) const data = JSON.parse(chunk)
const deltaText = responseText.substring(splitIndexEnd) updateChat(
try { +uuid,
const data = JSON.parse(firstChunk) index,
if (firstTime) { {
firstTime = false dateTime: new Date().toLocaleString(),
renderText = data.text ?? '' text: lastText + data.text ?? '',
} inversion: false,
else { error: false,
renderText = deltaText ?? '' loading: false,
} conversationOptions: { conversationId: data.conversationId, parentMessageId: data.id },
updateChat( requestOptions: { prompt: message, ...options },
+uuid, },
index, )
{
dateTime: new Date().toLocaleString(), if (openLongReply && data.detail.choices[0].finish_reason === 'length') {
text: renderText, options.parentMessageId = data.id
inversion: false, lastText = data.text
error: false, message = ''
loading: false, return fetchChatAPIOnce()
conversationOptions: { conversationId: data.conversationId, parentMessageId: data.id },
requestOptions: { prompt: message, ...options },
},
)
if (openLongReply && data.detail.choices[0].finish_reason === 'length') {
options.parentMessageId = data.id
message = ''
return fetchChatAPIOnce()
}
}
catch (error) {
//
} }
} }
catch (error) {
//
}
}, },
}) })
} }
@ -488,13 +467,20 @@ onUnmounted(() => {
<template> <template>
<div class="flex flex-col w-full h-full"> <div class="flex flex-col w-full h-full">
<HeaderComponent <HeaderComponent
v-if="isMobile" :using-context="usingContext" @export="handleExport" v-if="isMobile"
:using-context="usingContext"
@export="handleExport"
@toggle-using-context="toggleUsingContext" @toggle-using-context="toggleUsingContext"
/> />
<main class="flex-1 overflow-hidden"> <main class="flex-1 overflow-hidden">
<div id="scrollRef" ref="scrollRef" class="h-full overflow-hidden overflow-y-auto"> <div
id="scrollRef"
ref="scrollRef"
class="h-full overflow-hidden overflow-y-auto"
>
<div <div
id="image-wrapper" class="w-full max-w-screen-xl m-auto dark:bg-[#101014]" id="image-wrapper"
class="w-full max-w-screen-xl m-auto dark:bg-[#101014]"
:class="[isMobile ? 'p-2' : 'p-4']" :class="[isMobile ? 'p-2' : 'p-4']"
> >
<template v-if="!dataSources.length"> <template v-if="!dataSources.length">
@ -506,8 +492,14 @@ onUnmounted(() => {
<template v-else> <template v-else>
<div> <div>
<Message <Message
v-for="(item, index) of dataSources" :key="index" :date-time="item.dateTime" :text="item.text" v-for="(item, index) of dataSources"
:inversion="item.inversion" :error="item.error" :loading="item.loading" @regenerate="onRegenerate(index)" :key="index"
:date-time="item.dateTime"
:text="item.text"
:inversion="item.inversion"
:error="item.error"
:loading="item.loading"
@regenerate="onRegenerate(index)"
@delete="handleDelete(index)" @delete="handleDelete(index)"
/> />
<div class="sticky bottom-0 left-0 flex justify-center"> <div class="sticky bottom-0 left-0 flex justify-center">
@ -544,9 +536,15 @@ onUnmounted(() => {
<NAutoComplete v-model:value="prompt" :options="searchOptions" :render-label="renderOption"> <NAutoComplete v-model:value="prompt" :options="searchOptions" :render-label="renderOption">
<template #default="{ handleInput, handleBlur, handleFocus }"> <template #default="{ handleInput, handleBlur, handleFocus }">
<NInput <NInput
ref="inputRef" v-model:value="prompt" type="textarea" :placeholder="placeholder" ref="inputRef"
:autosize="{ minRows: 1, maxRows: isMobile ? 4 : 8 }" @input="handleInput" @focus="handleFocus" v-model:value="prompt"
@blur="handleBlur" @keypress="handleEnter" type="textarea"
:placeholder="placeholder"
:autosize="{ minRows: 1, maxRows: isMobile ? 4 : 8 }"
@input="handleInput"
@focus="handleFocus"
@blur="handleBlur"
@keypress="handleEnter"
/> />
</template> </template>
</NAutoComplete> </NAutoComplete>

Loading…
Cancel
Save