chat.ts 19 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653
  1. import { create } from "zustand";
  2. import { persist } from "zustand/middleware";
  3. import { trimTopic } from "../utils";
  4. import Locale, { getLang } from "../locales";
  5. import { showToast } from "../components/ui-lib";
  6. import { ModelConfig, ModelType, useAppConfig } from "./config";
  7. import { createEmptyMask, Mask } from "./mask";
  8. import {
  9. DEFAULT_INPUT_TEMPLATE,
  10. DEFAULT_SYSTEM_TEMPLATE,
  11. StoreKey,
  12. SUMMARIZE_MODEL,
  13. } from "../constant";
  14. import { api, RequestMessage } from "../client/api";
  15. import { ChatControllerPool } from "../client/controller";
  16. import { prettyObject } from "../utils/format";
  17. import { estimateTokenLength } from "../utils/token";
  18. import { nanoid } from "nanoid";
  19. export type ChatMessage = RequestMessage & {
  20. date: string;
  21. streaming?: boolean;
  22. isError?: boolean;
  23. id: string;
  24. model?: ModelType;
  25. };
  26. export function createMessage(override: Partial<ChatMessage>): ChatMessage {
  27. return {
  28. id: nanoid(),
  29. date: new Date().toLocaleString(),
  30. role: "user",
  31. content: "",
  32. ...override,
  33. };
  34. }
  35. export interface ChatStat {
  36. tokenCount: number;
  37. wordCount: number;
  38. charCount: number;
  39. }
  40. export interface ChatSession {
  41. id: string;
  42. topic: string;
  43. memoryPrompt: string;
  44. messages: ChatMessage[];
  45. stat: ChatStat;
  46. lastUpdate: number;
  47. lastSummarizeIndex: number;
  48. clearContextIndex?: number;
  49. mask: Mask;
  50. }
  51. export const DEFAULT_TOPIC = Locale.Store.DefaultTopic;
  52. export const BOT_HELLO: ChatMessage = createMessage({
  53. role: "assistant",
  54. content: Locale.Store.BotHello,
  55. });
  56. function createEmptySession(): ChatSession {
  57. return {
  58. id: nanoid(),
  59. topic: DEFAULT_TOPIC,
  60. memoryPrompt: "",
  61. messages: [],
  62. stat: {
  63. tokenCount: 0,
  64. wordCount: 0,
  65. charCount: 0,
  66. },
  67. lastUpdate: Date.now(),
  68. lastSummarizeIndex: 0,
  69. mask: createEmptyMask(),
  70. };
  71. }
  72. function getSummarizeModel(currentModel: string) {
  73. // if it is using gpt-* models, force to use 3.5 to summarize
  74. return currentModel.startsWith("gpt") ? SUMMARIZE_MODEL : currentModel;
  75. }
  76. interface ChatStore {
  77. sessions: ChatSession[];
  78. currentSessionIndex: number;
  79. clearSessions: () => void;
  80. moveSession: (from: number, to: number) => void;
  81. selectSession: (index: number) => void;
  82. newSession: (mask?: Mask) => void;
  83. deleteSession: (index: number) => void;
  84. currentSession: () => ChatSession;
  85. nextSession: (delta: number) => void;
  86. onNewMessage: (message: ChatMessage) => void;
  87. onUserInput: (content: string) => Promise<void>;
  88. summarizeSession: () => void;
  89. updateStat: (message: ChatMessage) => void;
  90. updateCurrentSession: (updater: (session: ChatSession) => void) => void;
  91. updateMessage: (
  92. sessionIndex: number,
  93. messageIndex: number,
  94. updater: (message?: ChatMessage) => void,
  95. ) => void;
  96. resetSession: () => void;
  97. getMessagesWithMemory: () => ChatMessage[];
  98. getMemoryPrompt: () => ChatMessage;
  99. clearAllData: () => void;
  100. }
  101. function countMessages(msgs: ChatMessage[]) {
  102. return msgs.reduce((pre, cur) => pre + estimateTokenLength(cur.content), 0);
  103. }
  104. function fillTemplateWith(input: string, modelConfig: ModelConfig) {
  105. const vars = {
  106. model: modelConfig.model,
  107. time: new Date().toLocaleString(),
  108. lang: getLang(),
  109. input: input,
  110. };
  111. let output = modelConfig.template ?? DEFAULT_INPUT_TEMPLATE;
  112. // must contains {{input}}
  113. const inputVar = "{{input}}";
  114. if (!output.includes(inputVar)) {
  115. output += "\n" + inputVar;
  116. }
  117. Object.entries(vars).forEach(([name, value]) => {
  118. output = output.replaceAll(`{{${name}}}`, value);
  119. });
  120. return output;
  121. }
  122. export const useChatStore = create<ChatStore>()(
  123. persist(
  124. (set, get) => ({
  125. sessions: [createEmptySession()],
  126. currentSessionIndex: 0,
  127. clearSessions() {
  128. set(() => ({
  129. sessions: [createEmptySession()],
  130. currentSessionIndex: 0,
  131. }));
  132. },
  133. selectSession(index: number) {
  134. set({
  135. currentSessionIndex: index,
  136. });
  137. },
  138. moveSession(from: number, to: number) {
  139. set((state) => {
  140. const { sessions, currentSessionIndex: oldIndex } = state;
  141. // move the session
  142. const newSessions = [...sessions];
  143. const session = newSessions[from];
  144. newSessions.splice(from, 1);
  145. newSessions.splice(to, 0, session);
  146. // modify current session id
  147. let newIndex = oldIndex === from ? to : oldIndex;
  148. if (oldIndex > from && oldIndex <= to) {
  149. newIndex -= 1;
  150. } else if (oldIndex < from && oldIndex >= to) {
  151. newIndex += 1;
  152. }
  153. return {
  154. currentSessionIndex: newIndex,
  155. sessions: newSessions,
  156. };
  157. });
  158. },
  159. newSession(mask) {
  160. const session = createEmptySession();
  161. if (mask) {
  162. const config = useAppConfig.getState();
  163. const globalModelConfig = config.modelConfig;
  164. session.mask = {
  165. ...mask,
  166. modelConfig: {
  167. ...globalModelConfig,
  168. ...mask.modelConfig,
  169. },
  170. };
  171. session.topic = mask.name;
  172. }
  173. set((state) => ({
  174. currentSessionIndex: 0,
  175. sessions: [session].concat(state.sessions),
  176. }));
  177. },
  178. nextSession(delta) {
  179. const n = get().sessions.length;
  180. const limit = (x: number) => (x + n) % n;
  181. const i = get().currentSessionIndex;
  182. get().selectSession(limit(i + delta));
  183. },
  184. deleteSession(index) {
  185. const deletingLastSession = get().sessions.length === 1;
  186. const deletedSession = get().sessions.at(index);
  187. if (!deletedSession) return;
  188. const sessions = get().sessions.slice();
  189. sessions.splice(index, 1);
  190. const currentIndex = get().currentSessionIndex;
  191. let nextIndex = Math.min(
  192. currentIndex - Number(index < currentIndex),
  193. sessions.length - 1,
  194. );
  195. if (deletingLastSession) {
  196. nextIndex = 0;
  197. sessions.push(createEmptySession());
  198. }
  199. // for undo delete action
  200. const restoreState = {
  201. currentSessionIndex: get().currentSessionIndex,
  202. sessions: get().sessions.slice(),
  203. };
  204. set(() => ({
  205. currentSessionIndex: nextIndex,
  206. sessions,
  207. }));
  208. showToast(
  209. Locale.Home.DeleteToast,
  210. {
  211. text: Locale.Home.Revert,
  212. onClick() {
  213. set(() => restoreState);
  214. },
  215. },
  216. 5000,
  217. );
  218. },
  219. currentSession() {
  220. let index = get().currentSessionIndex;
  221. const sessions = get().sessions;
  222. if (index < 0 || index >= sessions.length) {
  223. index = Math.min(sessions.length - 1, Math.max(0, index));
  224. set(() => ({ currentSessionIndex: index }));
  225. }
  226. const session = sessions[index];
  227. return session;
  228. },
  229. onNewMessage(message) {
  230. get().updateCurrentSession((session) => {
  231. session.messages = session.messages.concat();
  232. session.lastUpdate = Date.now();
  233. });
  234. get().updateStat(message);
  235. get().summarizeSession();
  236. },
  237. async onUserInput(content) {
  238. const session = get().currentSession();
  239. const modelConfig = session.mask.modelConfig;
  240. const userContent = fillTemplateWith(content, modelConfig);
  241. console.log("[User Input] after template: ", userContent);
  242. const userMessage: ChatMessage = createMessage({
  243. role: "user",
  244. content: userContent,
  245. });
  246. const botMessage: ChatMessage = createMessage({
  247. role: "assistant",
  248. streaming: true,
  249. model: modelConfig.model,
  250. });
  251. // get recent messages
  252. const recentMessages = get().getMessagesWithMemory();
  253. const sendMessages = recentMessages.concat(userMessage);
  254. const messageIndex = get().currentSession().messages.length + 1;
  255. // save user's and bot's message
  256. get().updateCurrentSession((session) => {
  257. const savedUserMessage = {
  258. ...userMessage,
  259. content,
  260. };
  261. session.messages = session.messages.concat([
  262. savedUserMessage,
  263. botMessage,
  264. ]);
  265. });
  266. // make request
  267. api.llm.chat({
  268. messages: sendMessages,
  269. config: { ...modelConfig, stream: true },
  270. onUpdate(message) {
  271. botMessage.streaming = true;
  272. if (message) {
  273. botMessage.content = message;
  274. }
  275. get().updateCurrentSession((session) => {
  276. session.messages = session.messages.concat();
  277. });
  278. },
  279. onFinish(message) {
  280. botMessage.streaming = false;
  281. if (message) {
  282. botMessage.content = message;
  283. get().onNewMessage(botMessage);
  284. }
  285. ChatControllerPool.remove(session.id, botMessage.id);
  286. },
  287. onError(error) {
  288. const isAborted = error.message.includes("aborted");
  289. botMessage.content +=
  290. "\n\n" +
  291. prettyObject({
  292. error: true,
  293. message: error.message,
  294. });
  295. botMessage.streaming = false;
  296. userMessage.isError = !isAborted;
  297. botMessage.isError = !isAborted;
  298. get().updateCurrentSession((session) => {
  299. session.messages = session.messages.concat();
  300. });
  301. ChatControllerPool.remove(
  302. session.id,
  303. botMessage.id ?? messageIndex,
  304. );
  305. console.error("[Chat] failed ", error);
  306. },
  307. onController(controller) {
  308. // collect controller for stop/retry
  309. ChatControllerPool.addController(
  310. session.id,
  311. botMessage.id ?? messageIndex,
  312. controller,
  313. );
  314. },
  315. });
  316. },
  317. getMemoryPrompt() {
  318. const session = get().currentSession();
  319. return {
  320. role: "system",
  321. content:
  322. session.memoryPrompt.length > 0
  323. ? Locale.Store.Prompt.History(session.memoryPrompt)
  324. : "",
  325. date: "",
  326. } as ChatMessage;
  327. },
  328. getMessagesWithMemory() {
  329. const session = get().currentSession();
  330. const modelConfig = session.mask.modelConfig;
  331. const clearContextIndex = session.clearContextIndex ?? 0;
  332. const messages = session.messages.slice();
  333. const totalMessageCount = session.messages.length;
  334. // in-context prompts
  335. const contextPrompts = session.mask.context.slice();
  336. // system prompts, to get close to OpenAI Web ChatGPT
  337. const shouldInjectSystemPrompts = modelConfig.enableInjectSystemPrompts;
  338. const systemPrompts = shouldInjectSystemPrompts
  339. ? [
  340. createMessage({
  341. role: "system",
  342. content: fillTemplateWith("", {
  343. ...modelConfig,
  344. template: DEFAULT_SYSTEM_TEMPLATE,
  345. }),
  346. }),
  347. ]
  348. : [];
  349. if (shouldInjectSystemPrompts) {
  350. console.log(
  351. "[Global System Prompt] ",
  352. systemPrompts.at(0)?.content ?? "empty",
  353. );
  354. }
  355. // long term memory
  356. const shouldSendLongTermMemory =
  357. modelConfig.sendMemory &&
  358. session.memoryPrompt &&
  359. session.memoryPrompt.length > 0 &&
  360. session.lastSummarizeIndex > clearContextIndex;
  361. const longTermMemoryPrompts = shouldSendLongTermMemory
  362. ? [get().getMemoryPrompt()]
  363. : [];
  364. const longTermMemoryStartIndex = session.lastSummarizeIndex;
  365. // short term memory
  366. const shortTermMemoryStartIndex = Math.max(
  367. 0,
  368. totalMessageCount - modelConfig.historyMessageCount,
  369. );
  370. // lets concat send messages, including 4 parts:
  371. // 0. system prompt: to get close to OpenAI Web ChatGPT
  372. // 1. long term memory: summarized memory messages
  373. // 2. pre-defined in-context prompts
  374. // 3. short term memory: latest n messages
  375. // 4. newest input message
  376. const memoryStartIndex = shouldSendLongTermMemory
  377. ? Math.min(longTermMemoryStartIndex, shortTermMemoryStartIndex)
  378. : shortTermMemoryStartIndex;
  379. // and if user has cleared history messages, we should exclude the memory too.
  380. const contextStartIndex = Math.max(clearContextIndex, memoryStartIndex);
  381. const maxTokenThreshold = modelConfig.max_tokens;
  382. // get recent messages as much as possible
  383. const reversedRecentMessages = [];
  384. for (
  385. let i = totalMessageCount - 1, tokenCount = 0;
  386. i >= contextStartIndex && tokenCount < maxTokenThreshold;
  387. i -= 1
  388. ) {
  389. const msg = messages[i];
  390. if (!msg || msg.isError) continue;
  391. tokenCount += estimateTokenLength(msg.content);
  392. reversedRecentMessages.push(msg);
  393. }
  394. // concat all messages
  395. const recentMessages = [
  396. ...systemPrompts,
  397. ...longTermMemoryPrompts,
  398. ...contextPrompts,
  399. ...reversedRecentMessages.reverse(),
  400. ];
  401. return recentMessages;
  402. },
  403. updateMessage(
  404. sessionIndex: number,
  405. messageIndex: number,
  406. updater: (message?: ChatMessage) => void,
  407. ) {
  408. const sessions = get().sessions;
  409. const session = sessions.at(sessionIndex);
  410. const messages = session?.messages;
  411. updater(messages?.at(messageIndex));
  412. set(() => ({ sessions }));
  413. },
  414. resetSession() {
  415. get().updateCurrentSession((session) => {
  416. session.messages = [];
  417. session.memoryPrompt = "";
  418. });
  419. },
  420. summarizeSession() {
  421. const config = useAppConfig.getState();
  422. const session = get().currentSession();
  423. // remove error messages if any
  424. const messages = session.messages;
  425. // should summarize topic after chating more than 50 words
  426. const SUMMARIZE_MIN_LEN = 50;
  427. if (
  428. config.enableAutoGenerateTitle &&
  429. session.topic === DEFAULT_TOPIC &&
  430. countMessages(messages) >= SUMMARIZE_MIN_LEN
  431. ) {
  432. const topicMessages = messages.concat(
  433. createMessage({
  434. role: "user",
  435. content: Locale.Store.Prompt.Topic,
  436. }),
  437. );
  438. api.llm.chat({
  439. messages: topicMessages,
  440. config: {
  441. model: getSummarizeModel(session.mask.modelConfig.model),
  442. },
  443. onFinish(message) {
  444. get().updateCurrentSession(
  445. (session) =>
  446. (session.topic =
  447. message.length > 0 ? trimTopic(message) : DEFAULT_TOPIC),
  448. );
  449. },
  450. });
  451. }
  452. const modelConfig = session.mask.modelConfig;
  453. const summarizeIndex = Math.max(
  454. session.lastSummarizeIndex,
  455. session.clearContextIndex ?? 0,
  456. );
  457. let toBeSummarizedMsgs = messages
  458. .filter((msg) => !msg.isError)
  459. .slice(summarizeIndex);
  460. const historyMsgLength = countMessages(toBeSummarizedMsgs);
  461. if (historyMsgLength > modelConfig?.max_tokens ?? 4000) {
  462. const n = toBeSummarizedMsgs.length;
  463. toBeSummarizedMsgs = toBeSummarizedMsgs.slice(
  464. Math.max(0, n - modelConfig.historyMessageCount),
  465. );
  466. }
  467. // add memory prompt
  468. toBeSummarizedMsgs.unshift(get().getMemoryPrompt());
  469. const lastSummarizeIndex = session.messages.length;
  470. console.log(
  471. "[Chat History] ",
  472. toBeSummarizedMsgs,
  473. historyMsgLength,
  474. modelConfig.compressMessageLengthThreshold,
  475. );
  476. if (
  477. historyMsgLength > modelConfig.compressMessageLengthThreshold &&
  478. modelConfig.sendMemory
  479. ) {
  480. api.llm.chat({
  481. messages: toBeSummarizedMsgs.concat(
  482. createMessage({
  483. role: "system",
  484. content: Locale.Store.Prompt.Summarize,
  485. date: "",
  486. }),
  487. ),
  488. config: {
  489. ...modelConfig,
  490. stream: true,
  491. model: getSummarizeModel(session.mask.modelConfig.model),
  492. },
  493. onUpdate(message) {
  494. session.memoryPrompt = message;
  495. },
  496. onFinish(message) {
  497. console.log("[Memory] ", message);
  498. session.lastSummarizeIndex = lastSummarizeIndex;
  499. },
  500. onError(err) {
  501. console.error("[Summarize] ", err);
  502. },
  503. });
  504. }
  505. },
  506. updateStat(message) {
  507. get().updateCurrentSession((session) => {
  508. session.stat.charCount += message.content.length;
  509. // TODO: should update chat count and word count
  510. });
  511. },
  512. updateCurrentSession(updater) {
  513. const sessions = get().sessions;
  514. const index = get().currentSessionIndex;
  515. updater(sessions[index]);
  516. set(() => ({ sessions }));
  517. },
  518. clearAllData() {
  519. localStorage.clear();
  520. location.reload();
  521. },
  522. }),
  523. {
  524. name: StoreKey.Chat,
  525. version: 3.1,
  526. migrate(persistedState, version) {
  527. const state = persistedState as any;
  528. const newState = JSON.parse(JSON.stringify(state)) as ChatStore;
  529. if (version < 2) {
  530. newState.sessions = [];
  531. const oldSessions = state.sessions;
  532. for (const oldSession of oldSessions) {
  533. const newSession = createEmptySession();
  534. newSession.topic = oldSession.topic;
  535. newSession.messages = [...oldSession.messages];
  536. newSession.mask.modelConfig.sendMemory = true;
  537. newSession.mask.modelConfig.historyMessageCount = 4;
  538. newSession.mask.modelConfig.compressMessageLengthThreshold = 1000;
  539. newState.sessions.push(newSession);
  540. }
  541. }
  542. if (version < 3) {
  543. // migrate id to nanoid
  544. newState.sessions.forEach((s) => {
  545. s.id = nanoid();
  546. s.messages.forEach((m) => (m.id = nanoid()));
  547. });
  548. }
  549. // Enable `enableInjectSystemPrompts` attribute for old sessions.
  550. // Resolve issue of old sessions not automatically enabling.
  551. if (version < 3.1) {
  552. newState.sessions.forEach((s) => {
  553. if (
  554. // Exclude those already set by user
  555. !s.mask.modelConfig.hasOwnProperty("enableInjectSystemPrompts")
  556. ) {
  557. // Because users may have changed this configuration,
  558. // the user's current configuration is used instead of the default
  559. const config = useAppConfig.getState();
  560. s.mask.modelConfig.enableInjectSystemPrompts =
  561. config.modelConfig.enableInjectSystemPrompts;
  562. }
  563. });
  564. }
  565. return newState;
  566. },
  567. },
  568. ),
  569. );