requests.ts 6.6 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283
  1. import type { ChatRequest, ChatResponse } from "./api/openai/typing";
  2. import {
  3. Message,
  4. ModelConfig,
  5. ModelType,
  6. useAccessStore,
  7. useChatStore,
  8. } from "./store";
  9. import { showToast } from "./components/ui-lib";
  10. const TIME_OUT_MS = 60000;
  11. const makeRequestParam = (
  12. messages: Message[],
  13. options?: {
  14. filterBot?: boolean;
  15. stream?: boolean;
  16. model?: ModelType;
  17. },
  18. ): ChatRequest => {
  19. let sendMessages = messages.map((v) => ({
  20. role: v.role,
  21. content: v.content,
  22. }));
  23. if (options?.filterBot) {
  24. sendMessages = sendMessages.filter((m) => m.role !== "assistant");
  25. }
  26. const modelConfig = { ...useChatStore.getState().config.modelConfig };
  27. // @yidadaa: wont send max_tokens, because it is nonsense for Muggles
  28. // @ts-expect-error
  29. delete modelConfig.max_tokens;
  30. // override model config
  31. if (options?.model) {
  32. modelConfig.model = options.model;
  33. }
  34. return {
  35. messages: sendMessages,
  36. stream: options?.stream,
  37. ...modelConfig,
  38. };
  39. };
  40. function getHeaders() {
  41. const accessStore = useAccessStore.getState();
  42. let headers: Record<string, string> = {};
  43. if (accessStore.enabledAccessControl()) {
  44. headers["access-code"] = accessStore.accessCode;
  45. }
  46. if (accessStore.token && accessStore.token.length > 0) {
  47. headers["token"] = accessStore.token;
  48. }
  49. return headers;
  50. }
  51. export function requestOpenaiClient(path: string) {
  52. return (body: any, method = "POST") =>
  53. fetch("/api/openai", {
  54. method,
  55. headers: {
  56. "Content-Type": "application/json",
  57. path,
  58. ...getHeaders(),
  59. },
  60. body: body && JSON.stringify(body),
  61. });
  62. }
  63. export async function requestChat(
  64. messages: Message[],
  65. options?: {
  66. model?: ModelType;
  67. },
  68. ) {
  69. const req: ChatRequest = makeRequestParam(messages, {
  70. filterBot: true,
  71. model: options?.model,
  72. });
  73. const res = await requestOpenaiClient("v1/chat/completions")(req);
  74. try {
  75. const response = (await res.json()) as ChatResponse;
  76. return response;
  77. } catch (error) {
  78. console.error("[Request Chat] ", error, res.body);
  79. }
  80. }
  81. export async function requestUsage() {
  82. const formatDate = (d: Date) =>
  83. `${d.getFullYear()}-${(d.getMonth() + 1).toString().padStart(2, "0")}-${d
  84. .getDate()
  85. .toString()
  86. .padStart(2, "0")}`;
  87. const ONE_DAY = 2 * 24 * 60 * 60 * 1000;
  88. const now = new Date(Date.now() + ONE_DAY);
  89. const startOfMonth = new Date(now.getFullYear(), now.getMonth(), 1);
  90. const startDate = formatDate(startOfMonth);
  91. const endDate = formatDate(now);
  92. const [used, subs] = await Promise.all([
  93. requestOpenaiClient(
  94. `dashboard/billing/usage?start_date=${startDate}&end_date=${endDate}`,
  95. )(null, "GET"),
  96. requestOpenaiClient("dashboard/billing/subscription")(null, "GET"),
  97. ]);
  98. const response = (await used.json()) as {
  99. total_usage?: number;
  100. error?: {
  101. type: string;
  102. message: string;
  103. };
  104. };
  105. const total = (await subs.json()) as {
  106. hard_limit_usd?: number;
  107. };
  108. if (response.error && response.error.type) {
  109. showToast(response.error.message);
  110. return;
  111. }
  112. if (response.total_usage) {
  113. response.total_usage = Math.round(response.total_usage) / 100;
  114. }
  115. if (total.hard_limit_usd) {
  116. total.hard_limit_usd = Math.round(total.hard_limit_usd * 100) / 100;
  117. }
  118. return {
  119. used: response.total_usage,
  120. subscription: total.hard_limit_usd,
  121. };
  122. }
  123. export async function requestChatStream(
  124. messages: Message[],
  125. options?: {
  126. filterBot?: boolean;
  127. modelConfig?: ModelConfig;
  128. onMessage: (message: string, done: boolean) => void;
  129. onError: (error: Error, statusCode?: number) => void;
  130. onController?: (controller: AbortController) => void;
  131. },
  132. ) {
  133. const req = makeRequestParam(messages, {
  134. stream: true,
  135. filterBot: options?.filterBot,
  136. });
  137. console.log("[Request] ", req);
  138. const controller = new AbortController();
  139. const reqTimeoutId = setTimeout(() => controller.abort(), TIME_OUT_MS);
  140. try {
  141. const res = await fetch("/api/chat-stream", {
  142. method: "POST",
  143. headers: {
  144. "Content-Type": "application/json",
  145. path: "v1/chat/completions",
  146. ...getHeaders(),
  147. },
  148. body: JSON.stringify(req),
  149. signal: controller.signal,
  150. });
  151. clearTimeout(reqTimeoutId);
  152. let responseText = "";
  153. const finish = () => {
  154. options?.onMessage(responseText, true);
  155. controller.abort();
  156. };
  157. if (res.ok) {
  158. const reader = res.body?.getReader();
  159. const decoder = new TextDecoder();
  160. options?.onController?.(controller);
  161. while (true) {
  162. const resTimeoutId = setTimeout(() => finish(), TIME_OUT_MS);
  163. const content = await reader?.read();
  164. clearTimeout(resTimeoutId);
  165. if (!content || !content.value) {
  166. break;
  167. }
  168. const text = decoder.decode(content.value, { stream: true });
  169. responseText += text;
  170. const done = content.done;
  171. options?.onMessage(responseText, false);
  172. if (done) {
  173. break;
  174. }
  175. }
  176. finish();
  177. } else if (res.status === 401) {
  178. console.error("Unauthorized");
  179. options?.onError(new Error("Unauthorized"), res.status);
  180. } else {
  181. console.error("Stream Error", res.body);
  182. options?.onError(new Error("Stream Error"), res.status);
  183. }
  184. } catch (err) {
  185. console.error("NetWork Error", err);
  186. options?.onError(err as Error);
  187. }
  188. }
  189. export async function requestWithPrompt(
  190. messages: Message[],
  191. prompt: string,
  192. options?: {
  193. model?: ModelType;
  194. },
  195. ) {
  196. messages = messages.concat([
  197. {
  198. role: "user",
  199. content: prompt,
  200. date: new Date().toLocaleString(),
  201. },
  202. ]);
  203. const res = await requestChat(messages, options);
  204. return res?.choices?.at(0)?.message?.content ?? "";
  205. }
  206. // To store message streaming controller
  207. export const ControllerPool = {
  208. controllers: {} as Record<string, AbortController>,
  209. addController(
  210. sessionIndex: number,
  211. messageId: number,
  212. controller: AbortController,
  213. ) {
  214. const key = this.key(sessionIndex, messageId);
  215. this.controllers[key] = controller;
  216. return key;
  217. },
  218. stop(sessionIndex: number, messageId: number) {
  219. const key = this.key(sessionIndex, messageId);
  220. const controller = this.controllers[key];
  221. controller?.abort();
  222. },
  223. stopAll() {
  224. Object.values(this.controllers).forEach((v) => v.abort());
  225. },
  226. hasPending() {
  227. return Object.values(this.controllers).length > 0;
  228. },
  229. remove(sessionIndex: number, messageId: number) {
  230. const key = this.key(sessionIndex, messageId);
  231. delete this.controllers[key];
  232. },
  233. key(sessionIndex: number, messageIndex: number) {
  234. return `${sessionIndex},${messageIndex}`;
  235. },
  236. };