chat.ts 19 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668
  1. import { create } from "zustand";
  2. import { persist } from "zustand/middleware";
  3. import { trimTopic } from "../utils";
  4. import Locale, { getLang } from "../locales";
  5. import { showToast } from "../components/ui-lib";
  6. import { ModelConfig, ModelType, useAppConfig } from "./config";
  7. import { createEmptyMask, Mask } from "./mask";
  8. import {
  9. DEFAULT_INPUT_TEMPLATE,
  10. DEFAULT_SYSTEM_TEMPLATE,
  11. StoreKey,
  12. SUMMARIZE_MODEL,
  13. } from "../constant";
  14. import { api, RequestMessage } from "../client/api";
  15. import { ChatControllerPool } from "../client/controller";
  16. import { prettyObject } from "../utils/format";
  17. import { estimateTokenLength } from "../utils/token";
  18. import { nanoid } from "nanoid";
  19. import { createPersistStore } from "../utils/store";
  20. export type ChatMessage = RequestMessage & {
  21. date: string;
  22. streaming?: boolean;
  23. isError?: boolean;
  24. id: string;
  25. model?: ModelType;
  26. };
  27. export function createMessage(override: Partial<ChatMessage>): ChatMessage {
  28. return {
  29. id: nanoid(),
  30. date: new Date().toLocaleString(),
  31. role: "user",
  32. content: "",
  33. ...override,
  34. };
  35. }
  36. export interface ChatStat {
  37. tokenCount: number;
  38. wordCount: number;
  39. charCount: number;
  40. }
  41. export interface ChatSession {
  42. id: string;
  43. topic: string;
  44. memoryPrompt: string;
  45. messages: ChatMessage[];
  46. stat: ChatStat;
  47. lastUpdate: number;
  48. lastSummarizeIndex: number;
  49. clearContextIndex?: number;
  50. mask: Mask;
  51. }
  52. export const DEFAULT_TOPIC = Locale.Store.DefaultTopic;
  53. export const BOT_HELLO: ChatMessage = createMessage({
  54. role: "assistant",
  55. content: Locale.Store.BotHello,
  56. });
  57. function createEmptySession(): ChatSession {
  58. return {
  59. id: nanoid(),
  60. topic: DEFAULT_TOPIC,
  61. memoryPrompt: "",
  62. messages: [],
  63. stat: {
  64. tokenCount: 0,
  65. wordCount: 0,
  66. charCount: 0,
  67. },
  68. lastUpdate: Date.now(),
  69. lastSummarizeIndex: 0,
  70. mask: createEmptyMask(),
  71. };
  72. }
  73. function getSummarizeModel(currentModel: string) {
  74. // if it is using gpt-* models, force to use 3.5 to summarize
  75. return currentModel.startsWith("gpt") ? SUMMARIZE_MODEL : currentModel;
  76. }
  77. interface ChatStore {
  78. sessions: ChatSession[];
  79. currentSessionIndex: number;
  80. clearSessions: () => void;
  81. moveSession: (from: number, to: number) => void;
  82. selectSession: (index: number) => void;
  83. newSession: (mask?: Mask) => void;
  84. deleteSession: (index: number) => void;
  85. currentSession: () => ChatSession;
  86. nextSession: (delta: number) => void;
  87. onNewMessage: (message: ChatMessage) => void;
  88. onUserInput: (content: string) => Promise<void>;
  89. summarizeSession: () => void;
  90. updateStat: (message: ChatMessage) => void;
  91. updateCurrentSession: (updater: (session: ChatSession) => void) => void;
  92. updateMessage: (
  93. sessionIndex: number,
  94. messageIndex: number,
  95. updater: (message?: ChatMessage) => void,
  96. ) => void;
  97. resetSession: () => void;
  98. getMessagesWithMemory: () => ChatMessage[];
  99. getMemoryPrompt: () => ChatMessage;
  100. clearAllData: () => void;
  101. }
  102. function countMessages(msgs: ChatMessage[]) {
  103. return msgs.reduce((pre, cur) => pre + estimateTokenLength(cur.content), 0);
  104. }
  105. function fillTemplateWith(input: string, modelConfig: ModelConfig) {
  106. const vars = {
  107. model: modelConfig.model,
  108. time: new Date().toLocaleString(),
  109. lang: getLang(),
  110. input: input,
  111. };
  112. let output = modelConfig.template ?? DEFAULT_INPUT_TEMPLATE;
  113. // must contains {{input}}
  114. const inputVar = "{{input}}";
  115. if (!output.includes(inputVar)) {
  116. output += "\n" + inputVar;
  117. }
  118. Object.entries(vars).forEach(([name, value]) => {
  119. output = output.replaceAll(`{{${name}}}`, value);
  120. });
  121. return output;
  122. }
  123. const DEFAULT_CHAT_STATE = {
  124. sessions: [createEmptySession()],
  125. currentSessionIndex: 0,
  126. };
  127. export const useChatStore = createPersistStore(
  128. DEFAULT_CHAT_STATE,
  129. (set, _get) => {
  130. function get() {
  131. return {
  132. ..._get(),
  133. ...methods,
  134. };
  135. }
  136. const methods = {
  137. clearSessions() {
  138. set(() => ({
  139. sessions: [createEmptySession()],
  140. currentSessionIndex: 0,
  141. }));
  142. },
  143. selectSession(index: number) {
  144. set({
  145. currentSessionIndex: index,
  146. });
  147. },
  148. moveSession(from: number, to: number) {
  149. set((state) => {
  150. const { sessions, currentSessionIndex: oldIndex } = state;
  151. // move the session
  152. const newSessions = [...sessions];
  153. const session = newSessions[from];
  154. newSessions.splice(from, 1);
  155. newSessions.splice(to, 0, session);
  156. // modify current session id
  157. let newIndex = oldIndex === from ? to : oldIndex;
  158. if (oldIndex > from && oldIndex <= to) {
  159. newIndex -= 1;
  160. } else if (oldIndex < from && oldIndex >= to) {
  161. newIndex += 1;
  162. }
  163. return {
  164. currentSessionIndex: newIndex,
  165. sessions: newSessions,
  166. };
  167. });
  168. },
  169. newSession(mask: Mask) {
  170. const session = createEmptySession();
  171. if (mask) {
  172. const config = useAppConfig.getState();
  173. const globalModelConfig = config.modelConfig;
  174. session.mask = {
  175. ...mask,
  176. modelConfig: {
  177. ...globalModelConfig,
  178. ...mask.modelConfig,
  179. },
  180. };
  181. session.topic = mask.name;
  182. }
  183. set((state) => ({
  184. currentSessionIndex: 0,
  185. sessions: [session].concat(state.sessions),
  186. }));
  187. },
  188. nextSession(delta: number) {
  189. const n = get().sessions.length;
  190. const limit = (x: number) => (x + n) % n;
  191. const i = get().currentSessionIndex;
  192. get().selectSession(limit(i + delta));
  193. },
  194. deleteSession(index: number) {
  195. const deletingLastSession = get().sessions.length === 1;
  196. const deletedSession = get().sessions.at(index);
  197. if (!deletedSession) return;
  198. const sessions = get().sessions.slice();
  199. sessions.splice(index, 1);
  200. const currentIndex = get().currentSessionIndex;
  201. let nextIndex = Math.min(
  202. currentIndex - Number(index < currentIndex),
  203. sessions.length - 1,
  204. );
  205. if (deletingLastSession) {
  206. nextIndex = 0;
  207. sessions.push(createEmptySession());
  208. }
  209. // for undo delete action
  210. const restoreState = {
  211. currentSessionIndex: get().currentSessionIndex,
  212. sessions: get().sessions.slice(),
  213. };
  214. set(() => ({
  215. currentSessionIndex: nextIndex,
  216. sessions,
  217. }));
  218. showToast(
  219. Locale.Home.DeleteToast,
  220. {
  221. text: Locale.Home.Revert,
  222. onClick() {
  223. set(() => restoreState);
  224. },
  225. },
  226. 5000,
  227. );
  228. },
  229. currentSession() {
  230. let index = get().currentSessionIndex;
  231. const sessions = get().sessions;
  232. if (index < 0 || index >= sessions.length) {
  233. index = Math.min(sessions.length - 1, Math.max(0, index));
  234. set(() => ({ currentSessionIndex: index }));
  235. }
  236. const session = sessions[index];
  237. return session;
  238. },
  239. onNewMessage(message: ChatMessage) {
  240. get().updateCurrentSession((session) => {
  241. session.messages = session.messages.concat();
  242. session.lastUpdate = Date.now();
  243. });
  244. get().updateStat(message);
  245. get().summarizeSession();
  246. },
  247. async onUserInput(content: string) {
  248. const session = get().currentSession();
  249. const modelConfig = session.mask.modelConfig;
  250. const userContent = fillTemplateWith(content, modelConfig);
  251. console.log("[User Input] after template: ", userContent);
  252. const userMessage: ChatMessage = createMessage({
  253. role: "user",
  254. content: userContent,
  255. });
  256. const botMessage: ChatMessage = createMessage({
  257. role: "assistant",
  258. streaming: true,
  259. model: modelConfig.model,
  260. });
  261. // get recent messages
  262. const recentMessages = get().getMessagesWithMemory();
  263. const sendMessages = recentMessages.concat(userMessage);
  264. const messageIndex = get().currentSession().messages.length + 1;
  265. // save user's and bot's message
  266. get().updateCurrentSession((session) => {
  267. const savedUserMessage = {
  268. ...userMessage,
  269. content,
  270. };
  271. session.messages = session.messages.concat([
  272. savedUserMessage,
  273. botMessage,
  274. ]);
  275. });
  276. // make request
  277. api.llm.chat({
  278. messages: sendMessages,
  279. config: { ...modelConfig, stream: true },
  280. onUpdate(message) {
  281. botMessage.streaming = true;
  282. if (message) {
  283. botMessage.content = message;
  284. }
  285. get().updateCurrentSession((session) => {
  286. session.messages = session.messages.concat();
  287. });
  288. },
  289. onFinish(message) {
  290. botMessage.streaming = false;
  291. if (message) {
  292. botMessage.content = message;
  293. get().onNewMessage(botMessage);
  294. }
  295. ChatControllerPool.remove(session.id, botMessage.id);
  296. },
  297. onError(error) {
  298. const isAborted = error.message.includes("aborted");
  299. botMessage.content +=
  300. "\n\n" +
  301. prettyObject({
  302. error: true,
  303. message: error.message,
  304. });
  305. botMessage.streaming = false;
  306. userMessage.isError = !isAborted;
  307. botMessage.isError = !isAborted;
  308. get().updateCurrentSession((session) => {
  309. session.messages = session.messages.concat();
  310. });
  311. ChatControllerPool.remove(
  312. session.id,
  313. botMessage.id ?? messageIndex,
  314. );
  315. console.error("[Chat] failed ", error);
  316. },
  317. onController(controller) {
  318. // collect controller for stop/retry
  319. ChatControllerPool.addController(
  320. session.id,
  321. botMessage.id ?? messageIndex,
  322. controller,
  323. );
  324. },
  325. });
  326. },
  327. getMemoryPrompt() {
  328. const session = get().currentSession();
  329. return {
  330. role: "system",
  331. content:
  332. session.memoryPrompt.length > 0
  333. ? Locale.Store.Prompt.History(session.memoryPrompt)
  334. : "",
  335. date: "",
  336. } as ChatMessage;
  337. },
  338. getMessagesWithMemory() {
  339. const session = get().currentSession();
  340. const modelConfig = session.mask.modelConfig;
  341. const clearContextIndex = session.clearContextIndex ?? 0;
  342. const messages = session.messages.slice();
  343. const totalMessageCount = session.messages.length;
  344. // in-context prompts
  345. const contextPrompts = session.mask.context.slice();
  346. // system prompts, to get close to OpenAI Web ChatGPT
  347. const shouldInjectSystemPrompts = modelConfig.enableInjectSystemPrompts;
  348. const systemPrompts = shouldInjectSystemPrompts
  349. ? [
  350. createMessage({
  351. role: "system",
  352. content: fillTemplateWith("", {
  353. ...modelConfig,
  354. template: DEFAULT_SYSTEM_TEMPLATE,
  355. }),
  356. }),
  357. ]
  358. : [];
  359. if (shouldInjectSystemPrompts) {
  360. console.log(
  361. "[Global System Prompt] ",
  362. systemPrompts.at(0)?.content ?? "empty",
  363. );
  364. }
  365. // long term memory
  366. const shouldSendLongTermMemory =
  367. modelConfig.sendMemory &&
  368. session.memoryPrompt &&
  369. session.memoryPrompt.length > 0 &&
  370. session.lastSummarizeIndex > clearContextIndex;
  371. const longTermMemoryPrompts = shouldSendLongTermMemory
  372. ? [get().getMemoryPrompt()]
  373. : [];
  374. const longTermMemoryStartIndex = session.lastSummarizeIndex;
  375. // short term memory
  376. const shortTermMemoryStartIndex = Math.max(
  377. 0,
  378. totalMessageCount - modelConfig.historyMessageCount,
  379. );
  380. // lets concat send messages, including 4 parts:
  381. // 0. system prompt: to get close to OpenAI Web ChatGPT
  382. // 1. long term memory: summarized memory messages
  383. // 2. pre-defined in-context prompts
  384. // 3. short term memory: latest n messages
  385. // 4. newest input message
  386. const memoryStartIndex = shouldSendLongTermMemory
  387. ? Math.min(longTermMemoryStartIndex, shortTermMemoryStartIndex)
  388. : shortTermMemoryStartIndex;
  389. // and if user has cleared history messages, we should exclude the memory too.
  390. const contextStartIndex = Math.max(clearContextIndex, memoryStartIndex);
  391. const maxTokenThreshold = modelConfig.max_tokens;
  392. // get recent messages as much as possible
  393. const reversedRecentMessages = [];
  394. for (
  395. let i = totalMessageCount - 1, tokenCount = 0;
  396. i >= contextStartIndex && tokenCount < maxTokenThreshold;
  397. i -= 1
  398. ) {
  399. const msg = messages[i];
  400. if (!msg || msg.isError) continue;
  401. tokenCount += estimateTokenLength(msg.content);
  402. reversedRecentMessages.push(msg);
  403. }
  404. // concat all messages
  405. const recentMessages = [
  406. ...systemPrompts,
  407. ...longTermMemoryPrompts,
  408. ...contextPrompts,
  409. ...reversedRecentMessages.reverse(),
  410. ];
  411. return recentMessages;
  412. },
  413. updateMessage(
  414. sessionIndex: number,
  415. messageIndex: number,
  416. updater: (message?: ChatMessage) => void,
  417. ) {
  418. const sessions = get().sessions;
  419. const session = sessions.at(sessionIndex);
  420. const messages = session?.messages;
  421. updater(messages?.at(messageIndex));
  422. set(() => ({ sessions }));
  423. },
  424. resetSession() {
  425. get().updateCurrentSession((session) => {
  426. session.messages = [];
  427. session.memoryPrompt = "";
  428. });
  429. },
  430. summarizeSession() {
  431. const config = useAppConfig.getState();
  432. const session = get().currentSession();
  433. // remove error messages if any
  434. const messages = session.messages;
  435. // should summarize topic after chating more than 50 words
  436. const SUMMARIZE_MIN_LEN = 50;
  437. if (
  438. config.enableAutoGenerateTitle &&
  439. session.topic === DEFAULT_TOPIC &&
  440. countMessages(messages) >= SUMMARIZE_MIN_LEN
  441. ) {
  442. const topicMessages = messages.concat(
  443. createMessage({
  444. role: "user",
  445. content: Locale.Store.Prompt.Topic,
  446. }),
  447. );
  448. api.llm.chat({
  449. messages: topicMessages,
  450. config: {
  451. model: getSummarizeModel(session.mask.modelConfig.model),
  452. },
  453. onFinish(message) {
  454. get().updateCurrentSession(
  455. (session) =>
  456. (session.topic =
  457. message.length > 0 ? trimTopic(message) : DEFAULT_TOPIC),
  458. );
  459. },
  460. });
  461. }
  462. const modelConfig = session.mask.modelConfig;
  463. const summarizeIndex = Math.max(
  464. session.lastSummarizeIndex,
  465. session.clearContextIndex ?? 0,
  466. );
  467. let toBeSummarizedMsgs = messages
  468. .filter((msg) => !msg.isError)
  469. .slice(summarizeIndex);
  470. const historyMsgLength = countMessages(toBeSummarizedMsgs);
  471. if (historyMsgLength > modelConfig?.max_tokens ?? 4000) {
  472. const n = toBeSummarizedMsgs.length;
  473. toBeSummarizedMsgs = toBeSummarizedMsgs.slice(
  474. Math.max(0, n - modelConfig.historyMessageCount),
  475. );
  476. }
  477. // add memory prompt
  478. toBeSummarizedMsgs.unshift(get().getMemoryPrompt());
  479. const lastSummarizeIndex = session.messages.length;
  480. console.log(
  481. "[Chat History] ",
  482. toBeSummarizedMsgs,
  483. historyMsgLength,
  484. modelConfig.compressMessageLengthThreshold,
  485. );
  486. if (
  487. historyMsgLength > modelConfig.compressMessageLengthThreshold &&
  488. modelConfig.sendMemory
  489. ) {
  490. api.llm.chat({
  491. messages: toBeSummarizedMsgs.concat(
  492. createMessage({
  493. role: "system",
  494. content: Locale.Store.Prompt.Summarize,
  495. date: "",
  496. }),
  497. ),
  498. config: {
  499. ...modelConfig,
  500. stream: true,
  501. model: getSummarizeModel(session.mask.modelConfig.model),
  502. },
  503. onUpdate(message) {
  504. session.memoryPrompt = message;
  505. },
  506. onFinish(message) {
  507. console.log("[Memory] ", message);
  508. session.lastSummarizeIndex = lastSummarizeIndex;
  509. },
  510. onError(err) {
  511. console.error("[Summarize] ", err);
  512. },
  513. });
  514. }
  515. },
  516. updateStat(message: ChatMessage) {
  517. get().updateCurrentSession((session) => {
  518. session.stat.charCount += message.content.length;
  519. // TODO: should update chat count and word count
  520. });
  521. },
  522. updateCurrentSession(updater: (session: ChatSession) => void) {
  523. const sessions = get().sessions;
  524. const index = get().currentSessionIndex;
  525. updater(sessions[index]);
  526. set(() => ({ sessions }));
  527. },
  528. clearAllData() {
  529. localStorage.clear();
  530. location.reload();
  531. },
  532. };
  533. return methods;
  534. },
  535. {
  536. name: StoreKey.Chat,
  537. version: 3.1,
  538. migrate(persistedState, version) {
  539. const state = persistedState as any;
  540. const newState = JSON.parse(
  541. JSON.stringify(state),
  542. ) as typeof DEFAULT_CHAT_STATE;
  543. if (version < 2) {
  544. newState.sessions = [];
  545. const oldSessions = state.sessions;
  546. for (const oldSession of oldSessions) {
  547. const newSession = createEmptySession();
  548. newSession.topic = oldSession.topic;
  549. newSession.messages = [...oldSession.messages];
  550. newSession.mask.modelConfig.sendMemory = true;
  551. newSession.mask.modelConfig.historyMessageCount = 4;
  552. newSession.mask.modelConfig.compressMessageLengthThreshold = 1000;
  553. newState.sessions.push(newSession);
  554. }
  555. }
  556. if (version < 3) {
  557. // migrate id to nanoid
  558. newState.sessions.forEach((s) => {
  559. s.id = nanoid();
  560. s.messages.forEach((m) => (m.id = nanoid()));
  561. });
  562. }
  563. // Enable `enableInjectSystemPrompts` attribute for old sessions.
  564. // Resolve issue of old sessions not automatically enabling.
  565. if (version < 3.1) {
  566. newState.sessions.forEach((s) => {
  567. if (
  568. // Exclude those already set by user
  569. !s.mask.modelConfig.hasOwnProperty("enableInjectSystemPrompts")
  570. ) {
  571. // Because users may have changed this configuration,
  572. // the user's current configuration is used instead of the default
  573. const config = useAppConfig.getState();
  574. s.mask.modelConfig.enableInjectSystemPrompts =
  575. config.modelConfig.enableInjectSystemPrompts;
  576. }
  577. });
  578. }
  579. return newState as any;
  580. },
  581. },
  582. );