chat.ts 19 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669
  1. import { trimTopic } from "../utils";
  2. import Locale, { getLang } from "../locales";
  3. import { showToast } from "../components/ui-lib";
  4. import { ModelConfig, ModelType, useAppConfig } from "./config";
  5. import { createEmptyMask, Mask } from "./mask";
  6. import {
  7. DEFAULT_INPUT_TEMPLATE,
  8. DEFAULT_SYSTEM_TEMPLATE,
  9. StoreKey,
  10. SUMMARIZE_MODEL,
  11. } from "../constant";
  12. import { api, RequestMessage } from "../client/api";
  13. import { ChatControllerPool } from "../client/controller";
  14. import { prettyObject } from "../utils/format";
  15. import { estimateTokenLength } from "../utils/token";
  16. import { nanoid } from "nanoid";
  17. import { createPersistStore } from "../utils/store";
  18. export type ChatMessage = RequestMessage & {
  19. date: string;
  20. streaming?: boolean;
  21. isError?: boolean;
  22. id: string;
  23. model?: ModelType;
  24. };
  25. export function createMessage(override: Partial<ChatMessage>): ChatMessage {
  26. return {
  27. id: nanoid(),
  28. date: new Date().toLocaleString(),
  29. role: "user",
  30. content: "",
  31. ...override,
  32. };
  33. }
  34. export interface ChatStat {
  35. tokenCount: number;
  36. wordCount: number;
  37. charCount: number;
  38. }
  39. export interface ChatSession {
  40. id: string;
  41. topic: string;
  42. memoryPrompt: string;
  43. messages: ChatMessage[];
  44. stat: ChatStat;
  45. lastUpdate: number;
  46. lastSummarizeIndex: number;
  47. clearContextIndex?: number;
  48. mask: Mask;
  49. }
  50. export const DEFAULT_TOPIC = Locale.Store.DefaultTopic;
  51. export const BOT_HELLO: ChatMessage = createMessage({
  52. role: "assistant",
  53. content: Locale.Store.BotHello,
  54. });
  55. function createEmptySession(): ChatSession {
  56. return {
  57. id: nanoid(),
  58. topic: DEFAULT_TOPIC,
  59. memoryPrompt: "",
  60. messages: [],
  61. stat: {
  62. tokenCount: 0,
  63. wordCount: 0,
  64. charCount: 0,
  65. },
  66. lastUpdate: Date.now(),
  67. lastSummarizeIndex: 0,
  68. mask: createEmptyMask(),
  69. };
  70. }
  71. function getSummarizeModel(currentModel: string) {
  72. // if it is using gpt-* models, force to use 3.5 to summarize
  73. return currentModel.startsWith("gpt") ? SUMMARIZE_MODEL : currentModel;
  74. }
  75. interface ChatStore {
  76. sessions: ChatSession[];
  77. currentSessionIndex: number;
  78. clearSessions: () => void;
  79. moveSession: (from: number, to: number) => void;
  80. selectSession: (index: number) => void;
  81. newSession: (mask?: Mask) => void;
  82. deleteSession: (index: number) => void;
  83. currentSession: () => ChatSession;
  84. nextSession: (delta: number) => void;
  85. onNewMessage: (message: ChatMessage) => void;
  86. onUserInput: (content: string) => Promise<void>;
  87. summarizeSession: () => void;
  88. updateStat: (message: ChatMessage) => void;
  89. updateCurrentSession: (updater: (session: ChatSession) => void) => void;
  90. updateMessage: (
  91. sessionIndex: number,
  92. messageIndex: number,
  93. updater: (message?: ChatMessage) => void,
  94. ) => void;
  95. resetSession: () => void;
  96. getMessagesWithMemory: () => ChatMessage[];
  97. getMemoryPrompt: () => ChatMessage;
  98. clearAllData: () => void;
  99. }
  100. function countMessages(msgs: ChatMessage[]) {
  101. return msgs.reduce((pre, cur) => pre + estimateTokenLength(cur.content), 0);
  102. }
  103. function fillTemplateWith(input: string, modelConfig: ModelConfig) {
  104. const vars = {
  105. model: modelConfig.model,
  106. time: new Date().toLocaleString(),
  107. lang: getLang(),
  108. input: input,
  109. };
  110. let output = modelConfig.template ?? DEFAULT_INPUT_TEMPLATE;
  111. // must contains {{input}}
  112. const inputVar = "{{input}}";
  113. if (!output.includes(inputVar)) {
  114. output += "\n" + inputVar;
  115. }
  116. Object.entries(vars).forEach(([name, value]) => {
  117. output = output.replaceAll(`{{${name}}}`, value);
  118. });
  119. return output;
  120. }
  121. const DEFAULT_CHAT_STATE = {
  122. sessions: [createEmptySession()],
  123. currentSessionIndex: 0,
  124. };
  125. export const useChatStore = createPersistStore(
  126. DEFAULT_CHAT_STATE,
  127. (set, _get) => {
  128. function get() {
  129. return {
  130. ..._get(),
  131. ...methods,
  132. };
  133. }
  134. const methods = {
  135. clearSessions() {
  136. set(() => ({
  137. sessions: [createEmptySession()],
  138. currentSessionIndex: 0,
  139. }));
  140. },
  141. selectSession(index: number) {
  142. set({
  143. currentSessionIndex: index,
  144. });
  145. },
  146. moveSession(from: number, to: number) {
  147. set((state) => {
  148. const { sessions, currentSessionIndex: oldIndex } = state;
  149. // move the session
  150. const newSessions = [...sessions];
  151. const session = newSessions[from];
  152. newSessions.splice(from, 1);
  153. newSessions.splice(to, 0, session);
  154. // modify current session id
  155. let newIndex = oldIndex === from ? to : oldIndex;
  156. if (oldIndex > from && oldIndex <= to) {
  157. newIndex -= 1;
  158. } else if (oldIndex < from && oldIndex >= to) {
  159. newIndex += 1;
  160. }
  161. return {
  162. currentSessionIndex: newIndex,
  163. sessions: newSessions,
  164. };
  165. });
  166. },
  167. newSession(mask?: Mask) {
  168. const session = createEmptySession();
  169. if (mask) {
  170. const config = useAppConfig.getState();
  171. const globalModelConfig = config.modelConfig;
  172. session.mask = {
  173. ...mask,
  174. modelConfig: {
  175. ...globalModelConfig,
  176. ...mask.modelConfig,
  177. },
  178. };
  179. session.topic = mask.name;
  180. }
  181. set((state) => ({
  182. currentSessionIndex: 0,
  183. sessions: [session].concat(state.sessions),
  184. }));
  185. },
  186. nextSession(delta: number) {
  187. const n = get().sessions.length;
  188. const limit = (x: number) => (x + n) % n;
  189. const i = get().currentSessionIndex;
  190. get().selectSession(limit(i + delta));
  191. },
  192. deleteSession(index: number) {
  193. const deletingLastSession = get().sessions.length === 1;
  194. const deletedSession = get().sessions.at(index);
  195. if (!deletedSession) return;
  196. const sessions = get().sessions.slice();
  197. sessions.splice(index, 1);
  198. const currentIndex = get().currentSessionIndex;
  199. let nextIndex = Math.min(
  200. currentIndex - Number(index < currentIndex),
  201. sessions.length - 1,
  202. );
  203. if (deletingLastSession) {
  204. nextIndex = 0;
  205. sessions.push(createEmptySession());
  206. }
  207. // for undo delete action
  208. const restoreState = {
  209. currentSessionIndex: get().currentSessionIndex,
  210. sessions: get().sessions.slice(),
  211. };
  212. set(() => ({
  213. currentSessionIndex: nextIndex,
  214. sessions,
  215. }));
  216. showToast(
  217. Locale.Home.DeleteToast,
  218. {
  219. text: Locale.Home.Revert,
  220. onClick() {
  221. set(() => restoreState);
  222. },
  223. },
  224. 5000,
  225. );
  226. },
  227. currentSession() {
  228. let index = get().currentSessionIndex;
  229. const sessions = get().sessions;
  230. if (index < 0 || index >= sessions.length) {
  231. index = Math.min(sessions.length - 1, Math.max(0, index));
  232. set(() => ({ currentSessionIndex: index }));
  233. }
  234. const session = sessions[index];
  235. return session;
  236. },
  237. onNewMessage(message: ChatMessage) {
  238. get().updateCurrentSession((session) => {
  239. session.messages = session.messages.concat();
  240. session.lastUpdate = Date.now();
  241. });
  242. get().updateStat(message);
  243. get().summarizeSession();
  244. },
  245. async onUserInput(content: string) {
  246. const session = get().currentSession();
  247. const modelConfig = session.mask.modelConfig;
  248. const userContent = fillTemplateWith(content, modelConfig);
  249. console.log("[User Input] after template: ", userContent);
  250. const userMessage: ChatMessage = createMessage({
  251. role: "user",
  252. content: userContent,
  253. });
  254. const botMessage: ChatMessage = createMessage({
  255. role: "assistant",
  256. streaming: true,
  257. model: modelConfig.model,
  258. });
  259. // get recent messages
  260. const recentMessages = get().getMessagesWithMemory();
  261. const sendMessages = recentMessages.concat(userMessage);
  262. const messageIndex = get().currentSession().messages.length + 1;
  263. // save user's and bot's message
  264. get().updateCurrentSession((session) => {
  265. const savedUserMessage = {
  266. ...userMessage,
  267. content,
  268. };
  269. session.messages = session.messages.concat([
  270. savedUserMessage,
  271. botMessage,
  272. ]);
  273. });
  274. // make request
  275. api.llm.chat({
  276. messages: sendMessages,
  277. config: { ...modelConfig, stream: true },
  278. onUpdate(message) {
  279. botMessage.streaming = true;
  280. if (message) {
  281. botMessage.content = message;
  282. }
  283. get().updateCurrentSession((session) => {
  284. session.messages = session.messages.concat();
  285. });
  286. },
  287. onFinish(message) {
  288. botMessage.streaming = false;
  289. if (message) {
  290. botMessage.content = message;
  291. get().onNewMessage(botMessage);
  292. }
  293. ChatControllerPool.remove(session.id, botMessage.id);
  294. },
  295. onError(error) {
  296. const isAborted = error.message.includes("aborted");
  297. botMessage.content +=
  298. "\n\n" +
  299. prettyObject({
  300. error: true,
  301. message: error.message,
  302. });
  303. botMessage.streaming = false;
  304. userMessage.isError = !isAborted;
  305. botMessage.isError = !isAborted;
  306. get().updateCurrentSession((session) => {
  307. session.messages = session.messages.concat();
  308. });
  309. ChatControllerPool.remove(
  310. session.id,
  311. botMessage.id ?? messageIndex,
  312. );
  313. console.error("[Chat] failed ", error);
  314. },
  315. onController(controller) {
  316. // collect controller for stop/retry
  317. ChatControllerPool.addController(
  318. session.id,
  319. botMessage.id ?? messageIndex,
  320. controller,
  321. );
  322. },
  323. });
  324. },
  325. getMemoryPrompt() {
  326. const session = get().currentSession();
  327. return {
  328. role: "system",
  329. content:
  330. session.memoryPrompt.length > 0
  331. ? Locale.Store.Prompt.History(session.memoryPrompt)
  332. : "",
  333. date: "",
  334. } as ChatMessage;
  335. },
  336. getMessagesWithMemory() {
  337. const session = get().currentSession();
  338. const modelConfig = session.mask.modelConfig;
  339. const clearContextIndex = session.clearContextIndex ?? 0;
  340. const messages = session.messages.slice();
  341. const totalMessageCount = session.messages.length;
  342. // in-context prompts
  343. const contextPrompts = session.mask.context.slice();
  344. // system prompts, to get close to OpenAI Web ChatGPT
  345. const shouldInjectSystemPrompts = modelConfig.enableInjectSystemPrompts;
  346. let systemPrompts = shouldInjectSystemPrompts ? [] : [];
  347. if (shouldInjectSystemPrompts) {
  348. const model = modelConfig.model;
  349. let systemTemplate = DEFAULT_SYSTEM_TEMPLATE;
  350. if (model === "gpt-4-1106-preview" || model === "gpt-4-vision-preview") {
  351. systemTemplate = systemTemplate.replace("{{knowledgeCutoff}}", "2023-04");
  352. } else {
  353. systemTemplate = systemTemplate.replace("{{knowledgeCutoff}}", "2021-09");
  354. }
  355. const systemPrompt = createMessage({
  356. role: "system",
  357. content: fillTemplateWith("", {
  358. ...modelConfig,
  359. template: systemTemplate,
  360. }),
  361. });
  362. console.log("[Global System Prompt] ", systemPrompt.content);
  363. }
  364. // long term memory
  365. const shouldSendLongTermMemory =
  366. modelConfig.sendMemory &&
  367. session.memoryPrompt &&
  368. session.memoryPrompt.length > 0 &&
  369. session.lastSummarizeIndex > clearContextIndex;
  370. const longTermMemoryPrompts = shouldSendLongTermMemory
  371. ? [get().getMemoryPrompt()]
  372. : [];
  373. const longTermMemoryStartIndex = session.lastSummarizeIndex;
  374. // short term memory
  375. const shortTermMemoryStartIndex = Math.max(
  376. 0,
  377. totalMessageCount - modelConfig.historyMessageCount,
  378. );
  379. // lets concat send messages, including 4 parts:
  380. // 0. system prompt: to get close to OpenAI Web ChatGPT
  381. // 1. long term memory: summarized memory messages
  382. // 2. pre-defined in-context prompts
  383. // 3. short term memory: latest n messages
  384. // 4. newest input message
  385. const memoryStartIndex = shouldSendLongTermMemory
  386. ? Math.min(longTermMemoryStartIndex, shortTermMemoryStartIndex)
  387. : shortTermMemoryStartIndex;
  388. // and if user has cleared history messages, we should exclude the memory too.
  389. const contextStartIndex = Math.max(clearContextIndex, memoryStartIndex);
  390. const maxTokenThreshold = modelConfig.max_tokens;
  391. // get recent messages as much as possible
  392. const reversedRecentMessages = [];
  393. for (
  394. let i = totalMessageCount - 1, tokenCount = 0;
  395. i >= contextStartIndex && tokenCount < maxTokenThreshold;
  396. i -= 1
  397. ) {
  398. const msg = messages[i];
  399. if (!msg || msg.isError) continue;
  400. tokenCount += estimateTokenLength(msg.content);
  401. reversedRecentMessages.push(msg);
  402. }
  403. // concat all messages
  404. const recentMessages = [
  405. ...systemPrompts,
  406. ...longTermMemoryPrompts,
  407. ...contextPrompts,
  408. ...reversedRecentMessages.reverse(),
  409. ];
  410. return recentMessages;
  411. },
  412. updateMessage(
  413. sessionIndex: number,
  414. messageIndex: number,
  415. updater: (message?: ChatMessage) => void,
  416. ) {
  417. const sessions = get().sessions;
  418. const session = sessions.at(sessionIndex);
  419. const messages = session?.messages;
  420. updater(messages?.at(messageIndex));
  421. set(() => ({ sessions }));
  422. },
  423. resetSession() {
  424. get().updateCurrentSession((session) => {
  425. session.messages = [];
  426. session.memoryPrompt = "";
  427. });
  428. },
  429. summarizeSession() {
  430. const config = useAppConfig.getState();
  431. const session = get().currentSession();
  432. // remove error messages if any
  433. const messages = session.messages;
  434. // should summarize topic after chating more than 50 words
  435. const SUMMARIZE_MIN_LEN = 50;
  436. if (
  437. config.enableAutoGenerateTitle &&
  438. session.topic === DEFAULT_TOPIC &&
  439. countMessages(messages) >= SUMMARIZE_MIN_LEN
  440. ) {
  441. const topicMessages = messages.concat(
  442. createMessage({
  443. role: "user",
  444. content: Locale.Store.Prompt.Topic,
  445. }),
  446. );
  447. api.llm.chat({
  448. messages: topicMessages,
  449. config: {
  450. model: getSummarizeModel(session.mask.modelConfig.model),
  451. },
  452. onFinish(message) {
  453. get().updateCurrentSession(
  454. (session) =>
  455. (session.topic =
  456. message.length > 0 ? trimTopic(message) : DEFAULT_TOPIC),
  457. );
  458. },
  459. });
  460. }
  461. const modelConfig = session.mask.modelConfig;
  462. const summarizeIndex = Math.max(
  463. session.lastSummarizeIndex,
  464. session.clearContextIndex ?? 0,
  465. );
  466. let toBeSummarizedMsgs = messages
  467. .filter((msg) => !msg.isError)
  468. .slice(summarizeIndex);
  469. const historyMsgLength = countMessages(toBeSummarizedMsgs);
  470. if (historyMsgLength > modelConfig?.max_tokens ?? 4000) {
  471. const n = toBeSummarizedMsgs.length;
  472. toBeSummarizedMsgs = toBeSummarizedMsgs.slice(
  473. Math.max(0, n - modelConfig.historyMessageCount),
  474. );
  475. }
  476. // add memory prompt
  477. toBeSummarizedMsgs.unshift(get().getMemoryPrompt());
  478. const lastSummarizeIndex = session.messages.length;
  479. console.log(
  480. "[Chat History] ",
  481. toBeSummarizedMsgs,
  482. historyMsgLength,
  483. modelConfig.compressMessageLengthThreshold,
  484. );
  485. if (
  486. historyMsgLength > modelConfig.compressMessageLengthThreshold &&
  487. modelConfig.sendMemory
  488. ) {
  489. api.llm.chat({
  490. messages: toBeSummarizedMsgs.concat(
  491. createMessage({
  492. role: "system",
  493. content: Locale.Store.Prompt.Summarize,
  494. date: "",
  495. }),
  496. ),
  497. config: {
  498. ...modelConfig,
  499. stream: true,
  500. model: getSummarizeModel(session.mask.modelConfig.model),
  501. },
  502. onUpdate(message) {
  503. session.memoryPrompt = message;
  504. },
  505. onFinish(message) {
  506. console.log("[Memory] ", message);
  507. session.lastSummarizeIndex = lastSummarizeIndex;
  508. },
  509. onError(err) {
  510. console.error("[Summarize] ", err);
  511. },
  512. });
  513. }
  514. },
  515. updateStat(message: ChatMessage) {
  516. get().updateCurrentSession((session) => {
  517. session.stat.charCount += message.content.length;
  518. // TODO: should update chat count and word count
  519. });
  520. },
  521. updateCurrentSession(updater: (session: ChatSession) => void) {
  522. const sessions = get().sessions;
  523. const index = get().currentSessionIndex;
  524. updater(sessions[index]);
  525. set(() => ({ sessions }));
  526. },
  527. clearAllData() {
  528. localStorage.clear();
  529. location.reload();
  530. },
  531. };
  532. return methods;
  533. },
  534. {
  535. name: StoreKey.Chat,
  536. version: 3.1,
  537. migrate(persistedState, version) {
  538. const state = persistedState as any;
  539. const newState = JSON.parse(
  540. JSON.stringify(state),
  541. ) as typeof DEFAULT_CHAT_STATE;
  542. if (version < 2) {
  543. newState.sessions = [];
  544. const oldSessions = state.sessions;
  545. for (const oldSession of oldSessions) {
  546. const newSession = createEmptySession();
  547. newSession.topic = oldSession.topic;
  548. newSession.messages = [...oldSession.messages];
  549. newSession.mask.modelConfig.sendMemory = true;
  550. newSession.mask.modelConfig.historyMessageCount = 4;
  551. newSession.mask.modelConfig.compressMessageLengthThreshold = 1000;
  552. newState.sessions.push(newSession);
  553. }
  554. }
  555. if (version < 3) {
  556. // migrate id to nanoid
  557. newState.sessions.forEach((s) => {
  558. s.id = nanoid();
  559. s.messages.forEach((m) => (m.id = nanoid()));
  560. });
  561. }
  562. // Enable `enableInjectSystemPrompts` attribute for old sessions.
  563. // Resolve issue of old sessions not automatically enabling.
  564. if (version < 3.1) {
  565. newState.sessions.forEach((s) => {
  566. if (
  567. // Exclude those already set by user
  568. !s.mask.modelConfig.hasOwnProperty("enableInjectSystemPrompts")
  569. ) {
  570. // Because users may have changed this configuration,
  571. // the user's current configuration is used instead of the default
  572. const config = useAppConfig.getState();
  573. s.mask.modelConfig.enableInjectSystemPrompts =
  574. config.modelConfig.enableInjectSystemPrompts;
  575. }
  576. });
  577. }
  578. return newState as any;
  579. },
  580. },
  581. );