mirror of https://github.com/jkjoy/sunpeiwen.git
514 lines
20 KiB
TypeScript
514 lines
20 KiB
TypeScript
|
import Keyv from 'keyv';
|
||
|
|
||
|
type Role = 'user' | 'assistant' | 'system';
|
||
|
type FetchFn = typeof fetch;
|
||
|
type ChatGPTAPIOptions = {
|
||
|
apiKey: string;
|
||
|
/** @defaultValue `'https://api.openai.com'` **/
|
||
|
apiBaseUrl?: string;
|
||
|
apiOrg?: string;
|
||
|
/** @defaultValue `false` **/
|
||
|
debug?: boolean;
|
||
|
completionParams?: Partial<Omit<openai.CreateChatCompletionRequest, 'messages' | 'n' | 'stream'>>;
|
||
|
systemMessage?: string;
|
||
|
/** @defaultValue `4096` **/
|
||
|
maxModelTokens?: number;
|
||
|
/** @defaultValue `1000` **/
|
||
|
maxResponseTokens?: number;
|
||
|
messageStore?: Keyv;
|
||
|
getMessageById?: GetMessageByIdFunction;
|
||
|
upsertMessage?: UpsertMessageFunction;
|
||
|
fetch?: FetchFn;
|
||
|
};
|
||
|
type SendMessageOptions = {
|
||
|
/** The name of a user in a multi-user chat. */
|
||
|
name?: string;
|
||
|
parentMessageId?: string;
|
||
|
conversationId?: string;
|
||
|
messageId?: string;
|
||
|
stream?: boolean;
|
||
|
systemMessage?: string;
|
||
|
timeoutMs?: number;
|
||
|
onProgress?: (partialResponse: ChatMessage) => void;
|
||
|
abortSignal?: AbortSignal;
|
||
|
completionParams?: Partial<Omit<openai.CreateChatCompletionRequest, 'messages' | 'n' | 'stream'>>;
|
||
|
};
|
||
|
type MessageActionType = 'next' | 'variant';
|
||
|
type SendMessageBrowserOptions = {
|
||
|
conversationId?: string;
|
||
|
parentMessageId?: string;
|
||
|
messageId?: string;
|
||
|
action?: MessageActionType;
|
||
|
timeoutMs?: number;
|
||
|
onProgress?: (partialResponse: ChatMessage) => void;
|
||
|
abortSignal?: AbortSignal;
|
||
|
};
|
||
|
interface ChatMessage {
|
||
|
id: string;
|
||
|
text: string;
|
||
|
role: Role;
|
||
|
name?: string;
|
||
|
delta?: string;
|
||
|
detail?: openai.CreateChatCompletionResponse | CreateChatCompletionStreamResponse;
|
||
|
parentMessageId?: string;
|
||
|
conversationId?: string;
|
||
|
}
|
||
|
declare class ChatGPTError extends Error {
|
||
|
statusCode?: number;
|
||
|
statusText?: string;
|
||
|
isFinal?: boolean;
|
||
|
accountId?: string;
|
||
|
}
|
||
|
/** Returns a chat message from a store by it's ID (or null if not found). */
|
||
|
type GetMessageByIdFunction = (id: string) => Promise<ChatMessage>;
|
||
|
/** Upserts a chat message to a store. */
|
||
|
type UpsertMessageFunction = (message: ChatMessage) => Promise<void>;
|
||
|
interface CreateChatCompletionStreamResponse extends openai.CreateChatCompletionDeltaResponse {
|
||
|
usage: CreateCompletionStreamResponseUsage;
|
||
|
}
|
||
|
interface CreateCompletionStreamResponseUsage extends openai.CreateCompletionResponseUsage {
|
||
|
estimated: true;
|
||
|
}
|
||
|
/**
|
||
|
* https://chat.openapi.com/backend-api/conversation
|
||
|
*/
|
||
|
type ConversationJSONBody = {
|
||
|
/**
|
||
|
* The action to take
|
||
|
*/
|
||
|
action: string;
|
||
|
/**
|
||
|
* The ID of the conversation
|
||
|
*/
|
||
|
conversation_id?: string;
|
||
|
/**
|
||
|
* Prompts to provide
|
||
|
*/
|
||
|
messages: Prompt[];
|
||
|
/**
|
||
|
* The model to use
|
||
|
*/
|
||
|
model: string;
|
||
|
/**
|
||
|
* The parent message ID
|
||
|
*/
|
||
|
parent_message_id: string;
|
||
|
};
|
||
|
type Prompt = {
|
||
|
/**
|
||
|
* The content of the prompt
|
||
|
*/
|
||
|
content: PromptContent;
|
||
|
/**
|
||
|
* The ID of the prompt
|
||
|
*/
|
||
|
id: string;
|
||
|
/**
|
||
|
* The role played in the prompt
|
||
|
*/
|
||
|
role: Role;
|
||
|
};
|
||
|
type ContentType = 'text';
|
||
|
type PromptContent = {
|
||
|
/**
|
||
|
* The content type of the prompt
|
||
|
*/
|
||
|
content_type: ContentType;
|
||
|
/**
|
||
|
* The parts to the prompt
|
||
|
*/
|
||
|
parts: string[];
|
||
|
};
|
||
|
type ConversationResponseEvent = {
|
||
|
message?: Message;
|
||
|
conversation_id?: string;
|
||
|
error?: string | null;
|
||
|
};
|
||
|
type Message = {
|
||
|
id: string;
|
||
|
content: MessageContent;
|
||
|
role: Role;
|
||
|
user: string | null;
|
||
|
create_time: string | null;
|
||
|
update_time: string | null;
|
||
|
end_turn: null;
|
||
|
weight: number;
|
||
|
recipient: string;
|
||
|
metadata: MessageMetadata;
|
||
|
};
|
||
|
type MessageContent = {
|
||
|
content_type: string;
|
||
|
parts: string[];
|
||
|
};
|
||
|
type MessageMetadata = any;
|
||
|
declare namespace openai {
|
||
|
interface CreateChatCompletionDeltaResponse {
|
||
|
id: string;
|
||
|
object: 'chat.completion.chunk';
|
||
|
created: number;
|
||
|
model: string;
|
||
|
choices: [
|
||
|
{
|
||
|
delta: {
|
||
|
role: Role;
|
||
|
content?: string;
|
||
|
};
|
||
|
index: number;
|
||
|
finish_reason: string | null;
|
||
|
}
|
||
|
];
|
||
|
}
|
||
|
/**
|
||
|
*
|
||
|
* @export
|
||
|
* @interface ChatCompletionRequestMessage
|
||
|
*/
|
||
|
interface ChatCompletionRequestMessage {
|
||
|
/**
|
||
|
* The role of the author of this message.
|
||
|
* @type {string}
|
||
|
* @memberof ChatCompletionRequestMessage
|
||
|
*/
|
||
|
role: ChatCompletionRequestMessageRoleEnum;
|
||
|
/**
|
||
|
* The contents of the message
|
||
|
* @type {string}
|
||
|
* @memberof ChatCompletionRequestMessage
|
||
|
*/
|
||
|
content: string;
|
||
|
/**
|
||
|
* The name of the user in a multi-user chat
|
||
|
* @type {string}
|
||
|
* @memberof ChatCompletionRequestMessage
|
||
|
*/
|
||
|
name?: string;
|
||
|
}
|
||
|
const ChatCompletionRequestMessageRoleEnum: {
|
||
|
readonly System: 'system';
|
||
|
readonly User: 'user';
|
||
|
readonly Assistant: 'assistant';
|
||
|
};
|
||
|
type ChatCompletionRequestMessageRoleEnum = (typeof ChatCompletionRequestMessageRoleEnum)[keyof typeof ChatCompletionRequestMessageRoleEnum];
|
||
|
/**
|
||
|
*
|
||
|
* @export
|
||
|
* @interface ChatCompletionResponseMessage
|
||
|
*/
|
||
|
interface ChatCompletionResponseMessage {
|
||
|
/**
|
||
|
* The role of the author of this message.
|
||
|
* @type {string}
|
||
|
* @memberof ChatCompletionResponseMessage
|
||
|
*/
|
||
|
role: ChatCompletionResponseMessageRoleEnum;
|
||
|
/**
|
||
|
* The contents of the message
|
||
|
* @type {string}
|
||
|
* @memberof ChatCompletionResponseMessage
|
||
|
*/
|
||
|
content: string;
|
||
|
}
|
||
|
const ChatCompletionResponseMessageRoleEnum: {
|
||
|
readonly System: 'system';
|
||
|
readonly User: 'user';
|
||
|
readonly Assistant: 'assistant';
|
||
|
};
|
||
|
type ChatCompletionResponseMessageRoleEnum = (typeof ChatCompletionResponseMessageRoleEnum)[keyof typeof ChatCompletionResponseMessageRoleEnum];
|
||
|
/**
|
||
|
*
|
||
|
* @export
|
||
|
* @interface CreateChatCompletionRequest
|
||
|
*/
|
||
|
interface CreateChatCompletionRequest {
|
||
|
/**
|
||
|
* ID of the model to use. Currently, only `gpt-3.5-turbo` and `gpt-3.5-turbo-0301` are supported.
|
||
|
* @type {string}
|
||
|
* @memberof CreateChatCompletionRequest
|
||
|
*/
|
||
|
model: string;
|
||
|
/**
|
||
|
* The messages to generate chat completions for, in the [chat format](/docs/guides/chat/introduction).
|
||
|
* @type {Array<ChatCompletionRequestMessage>}
|
||
|
* @memberof CreateChatCompletionRequest
|
||
|
*/
|
||
|
messages: Array<ChatCompletionRequestMessage>;
|
||
|
/**
|
||
|
* What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. We generally recommend altering this or `top_p` but not both.
|
||
|
* @type {number}
|
||
|
* @memberof CreateChatCompletionRequest
|
||
|
*/
|
||
|
temperature?: number | null;
|
||
|
/**
|
||
|
* An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered. We generally recommend altering this or `temperature` but not both.
|
||
|
* @type {number}
|
||
|
* @memberof CreateChatCompletionRequest
|
||
|
*/
|
||
|
top_p?: number | null;
|
||
|
/**
|
||
|
* How many chat completion choices to generate for each input message.
|
||
|
* @type {number}
|
||
|
* @memberof CreateChatCompletionRequest
|
||
|
*/
|
||
|
n?: number | null;
|
||
|
/**
|
||
|
* If set, partial message deltas will be sent, like in ChatGPT. Tokens will be sent as data-only [server-sent events](https://developer.mozilla.org/en-US/docs/Web/API/Server-sent_events/Using_server-sent_events#Event_stream_format) as they become available, with the stream terminated by a `data: [DONE]` message.
|
||
|
* @type {boolean}
|
||
|
* @memberof CreateChatCompletionRequest
|
||
|
*/
|
||
|
stream?: boolean | null;
|
||
|
/**
|
||
|
*
|
||
|
* @type {CreateChatCompletionRequestStop}
|
||
|
* @memberof CreateChatCompletionRequest
|
||
|
*/
|
||
|
stop?: CreateChatCompletionRequestStop;
|
||
|
/**
|
||
|
* The maximum number of tokens allowed for the generated answer. By default, the number of tokens the model can return will be (4096 - prompt tokens).
|
||
|
* @type {number}
|
||
|
* @memberof CreateChatCompletionRequest
|
||
|
*/
|
||
|
max_tokens?: number;
|
||
|
/**
|
||
|
* Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model\'s likelihood to talk about new topics. [See more information about frequency and presence penalties.](/docs/api-reference/parameter-details)
|
||
|
* @type {number}
|
||
|
* @memberof CreateChatCompletionRequest
|
||
|
*/
|
||
|
presence_penalty?: number | null;
|
||
|
/**
|
||
|
* Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model\'s likelihood to repeat the same line verbatim. [See more information about frequency and presence penalties.](/docs/api-reference/parameter-details)
|
||
|
* @type {number}
|
||
|
* @memberof CreateChatCompletionRequest
|
||
|
*/
|
||
|
frequency_penalty?: number | null;
|
||
|
/**
|
||
|
* Modify the likelihood of specified tokens appearing in the completion. Accepts a json object that maps tokens (specified by their token ID in the tokenizer) to an associated bias value from -100 to 100. Mathematically, the bias is added to the logits generated by the model prior to sampling. The exact effect will vary per model, but values between -1 and 1 should decrease or increase likelihood of selection; values like -100 or 100 should result in a ban or exclusive selection of the relevant token.
|
||
|
* @type {object}
|
||
|
* @memberof CreateChatCompletionRequest
|
||
|
*/
|
||
|
logit_bias?: object | null;
|
||
|
/**
|
||
|
* A unique identifier representing your end-user, which can help OpenAI to monitor and detect abuse. [Learn more](/docs/guides/safety-best-practices/end-user-ids).
|
||
|
* @type {string}
|
||
|
* @memberof CreateChatCompletionRequest
|
||
|
*/
|
||
|
user?: string;
|
||
|
}
|
||
|
/**
|
||
|
* @type CreateChatCompletionRequestStop
|
||
|
* Up to 4 sequences where the API will stop generating further tokens.
|
||
|
* @export
|
||
|
*/
|
||
|
type CreateChatCompletionRequestStop = Array<string> | string;
|
||
|
/**
|
||
|
*
|
||
|
* @export
|
||
|
* @interface CreateChatCompletionResponse
|
||
|
*/
|
||
|
interface CreateChatCompletionResponse {
|
||
|
/**
|
||
|
*
|
||
|
* @type {string}
|
||
|
* @memberof CreateChatCompletionResponse
|
||
|
*/
|
||
|
id: string;
|
||
|
/**
|
||
|
*
|
||
|
* @type {string}
|
||
|
* @memberof CreateChatCompletionResponse
|
||
|
*/
|
||
|
object: string;
|
||
|
/**
|
||
|
*
|
||
|
* @type {number}
|
||
|
* @memberof CreateChatCompletionResponse
|
||
|
*/
|
||
|
created: number;
|
||
|
/**
|
||
|
*
|
||
|
* @type {string}
|
||
|
* @memberof CreateChatCompletionResponse
|
||
|
*/
|
||
|
model: string;
|
||
|
/**
|
||
|
*
|
||
|
* @type {Array<CreateChatCompletionResponseChoicesInner>}
|
||
|
* @memberof CreateChatCompletionResponse
|
||
|
*/
|
||
|
choices: Array<CreateChatCompletionResponseChoicesInner>;
|
||
|
/**
|
||
|
*
|
||
|
* @type {CreateCompletionResponseUsage}
|
||
|
* @memberof CreateChatCompletionResponse
|
||
|
*/
|
||
|
usage?: CreateCompletionResponseUsage;
|
||
|
}
|
||
|
/**
|
||
|
*
|
||
|
* @export
|
||
|
* @interface CreateChatCompletionResponseChoicesInner
|
||
|
*/
|
||
|
interface CreateChatCompletionResponseChoicesInner {
|
||
|
/**
|
||
|
*
|
||
|
* @type {number}
|
||
|
* @memberof CreateChatCompletionResponseChoicesInner
|
||
|
*/
|
||
|
index?: number;
|
||
|
/**
|
||
|
*
|
||
|
* @type {ChatCompletionResponseMessage}
|
||
|
* @memberof CreateChatCompletionResponseChoicesInner
|
||
|
*/
|
||
|
message?: ChatCompletionResponseMessage;
|
||
|
/**
|
||
|
*
|
||
|
* @type {string}
|
||
|
* @memberof CreateChatCompletionResponseChoicesInner
|
||
|
*/
|
||
|
finish_reason?: string;
|
||
|
}
|
||
|
/**
|
||
|
*
|
||
|
* @export
|
||
|
* @interface CreateCompletionResponseUsage
|
||
|
*/
|
||
|
interface CreateCompletionResponseUsage {
|
||
|
/**
|
||
|
*
|
||
|
* @type {number}
|
||
|
* @memberof CreateCompletionResponseUsage
|
||
|
*/
|
||
|
prompt_tokens: number;
|
||
|
/**
|
||
|
*
|
||
|
* @type {number}
|
||
|
* @memberof CreateCompletionResponseUsage
|
||
|
*/
|
||
|
completion_tokens: number;
|
||
|
/**
|
||
|
*
|
||
|
* @type {number}
|
||
|
* @memberof CreateCompletionResponseUsage
|
||
|
*/
|
||
|
total_tokens: number;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
declare class ChatGPTAPI {
|
||
|
protected _apiKey: string;
|
||
|
protected _apiBaseUrl: string;
|
||
|
protected _apiOrg?: string;
|
||
|
protected _debug: boolean;
|
||
|
protected _systemMessage: string;
|
||
|
protected _completionParams: Omit<openai.CreateChatCompletionRequest, 'messages' | 'n'>;
|
||
|
protected _maxModelTokens: number;
|
||
|
protected _maxResponseTokens: number;
|
||
|
protected _fetch: FetchFn;
|
||
|
protected _getMessageById: GetMessageByIdFunction;
|
||
|
protected _upsertMessage: UpsertMessageFunction;
|
||
|
protected _messageStore: Keyv<ChatMessage>;
|
||
|
/**
|
||
|
* Creates a new client wrapper around OpenAI's chat completion API, mimicing the official ChatGPT webapp's functionality as closely as possible.
|
||
|
*
|
||
|
* @param apiKey - OpenAI API key (required).
|
||
|
* @param apiOrg - Optional OpenAI API organization (optional).
|
||
|
* @param apiBaseUrl - Optional override for the OpenAI API base URL.
|
||
|
* @param debug - Optional enables logging debugging info to stdout.
|
||
|
* @param completionParams - Param overrides to send to the [OpenAI chat completion API](https://platform.openai.com/docs/api-reference/chat/create). Options like `temperature` and `presence_penalty` can be tweaked to change the personality of the assistant.
|
||
|
* @param maxModelTokens - Optional override for the maximum number of tokens allowed by the model's context. Defaults to 4096.
|
||
|
* @param maxResponseTokens - Optional override for the minimum number of tokens allowed for the model's response. Defaults to 1000.
|
||
|
* @param messageStore - Optional [Keyv](https://github.com/jaredwray/keyv) store to persist chat messages to. If not provided, messages will be lost when the process exits.
|
||
|
* @param getMessageById - Optional function to retrieve a message by its ID. If not provided, the default implementation will be used (using an in-memory `messageStore`).
|
||
|
* @param upsertMessage - Optional function to insert or update a message. If not provided, the default implementation will be used (using an in-memory `messageStore`).
|
||
|
* @param fetch - Optional override for the `fetch` implementation to use. Defaults to the global `fetch` function.
|
||
|
*/
|
||
|
constructor(opts: ChatGPTAPIOptions);
|
||
|
/**
|
||
|
* Sends a message to the OpenAI chat completions endpoint, waits for the response
|
||
|
* to resolve, and returns the response.
|
||
|
*
|
||
|
* If you want your response to have historical context, you must provide a valid `parentMessageId`.
|
||
|
*
|
||
|
* If you want to receive a stream of partial responses, use `opts.onProgress`.
|
||
|
*
|
||
|
* Set `debug: true` in the `ChatGPTAPI` constructor to log more info on the full prompt sent to the OpenAI chat completions API. You can override the `systemMessage` in `opts` to customize the assistant's instructions.
|
||
|
*
|
||
|
* @param message - The prompt message to send
|
||
|
* @param opts.parentMessageId - Optional ID of the previous message in the conversation (defaults to `undefined`)
|
||
|
* @param opts.conversationId - Optional ID of the conversation (defaults to `undefined`)
|
||
|
* @param opts.messageId - Optional ID of the message to send (defaults to a random UUID)
|
||
|
* @param opts.systemMessage - Optional override for the chat "system message" which acts as instructions to the model (defaults to the ChatGPT system message)
|
||
|
* @param opts.timeoutMs - Optional timeout in milliseconds (defaults to no timeout)
|
||
|
* @param opts.onProgress - Optional callback which will be invoked every time the partial response is updated
|
||
|
* @param opts.abortSignal - Optional callback used to abort the underlying `fetch` call using an [AbortController](https://developer.mozilla.org/en-US/docs/Web/API/AbortController)
|
||
|
* @param completionParams - Optional overrides to send to the [OpenAI chat completion API](https://platform.openai.com/docs/api-reference/chat/create). Options like `temperature` and `presence_penalty` can be tweaked to change the personality of the assistant.
|
||
|
*
|
||
|
* @returns The response from ChatGPT
|
||
|
*/
|
||
|
sendMessage(text: string, opts?: SendMessageOptions): Promise<ChatMessage>;
|
||
|
get apiKey(): string;
|
||
|
set apiKey(apiKey: string);
|
||
|
get apiOrg(): string;
|
||
|
set apiOrg(apiOrg: string);
|
||
|
protected _buildMessages(text: string, opts: SendMessageOptions): Promise<{
|
||
|
messages: openai.ChatCompletionRequestMessage[];
|
||
|
maxTokens: number;
|
||
|
numTokens: number;
|
||
|
}>;
|
||
|
protected _getTokenCount(text: string): Promise<number>;
|
||
|
protected _defaultGetMessageById(id: string): Promise<ChatMessage>;
|
||
|
protected _defaultUpsertMessage(message: ChatMessage): Promise<void>;
|
||
|
}
|
||
|
|
||
|
declare class ChatGPTUnofficialProxyAPI {
|
||
|
protected _accessToken: string;
|
||
|
protected _apiReverseProxyUrl: string;
|
||
|
protected _debug: boolean;
|
||
|
protected _model: string;
|
||
|
protected _headers: Record<string, string>;
|
||
|
protected _fetch: FetchFn;
|
||
|
/**
|
||
|
* @param fetch - Optional override for the `fetch` implementation to use. Defaults to the global `fetch` function.
|
||
|
*/
|
||
|
constructor(opts: {
|
||
|
accessToken: string;
|
||
|
/** @defaultValue `https://bypass.duti.tech/api/conversation` **/
|
||
|
apiReverseProxyUrl?: string;
|
||
|
/** @defaultValue `text-davinci-002-render-sha` **/
|
||
|
model?: string;
|
||
|
/** @defaultValue `false` **/
|
||
|
debug?: boolean;
|
||
|
/** @defaultValue `undefined` **/
|
||
|
headers?: Record<string, string>;
|
||
|
fetch?: FetchFn;
|
||
|
});
|
||
|
get accessToken(): string;
|
||
|
set accessToken(value: string);
|
||
|
/**
|
||
|
* Sends a message to ChatGPT, waits for the response to resolve, and returns
|
||
|
* the response.
|
||
|
*
|
||
|
* If you want your response to have historical context, you must provide a valid `parentMessageId`.
|
||
|
*
|
||
|
* If you want to receive a stream of partial responses, use `opts.onProgress`.
|
||
|
* If you want to receive the full response, including message and conversation IDs,
|
||
|
* you can use `opts.onConversationResponse` or use the `ChatGPTAPI.getConversation`
|
||
|
* helper.
|
||
|
*
|
||
|
* Set `debug: true` in the `ChatGPTAPI` constructor to log more info on the full prompt sent to the OpenAI completions API. You can override the `promptPrefix` and `promptSuffix` in `opts` to customize the prompt.
|
||
|
*
|
||
|
* @param message - The prompt message to send
|
||
|
* @param opts.conversationId - Optional ID of a conversation to continue (defaults to a random UUID)
|
||
|
* @param opts.parentMessageId - Optional ID of the previous message in the conversation (defaults to `undefined`)
|
||
|
* @param opts.messageId - Optional ID of the message to send (defaults to a random UUID)
|
||
|
* @param opts.timeoutMs - Optional timeout in milliseconds (defaults to no timeout)
|
||
|
* @param opts.onProgress - Optional callback which will be invoked every time the partial response is updated
|
||
|
* @param opts.abortSignal - Optional callback used to abort the underlying `fetch` call using an [AbortController](https://developer.mozilla.org/en-US/docs/Web/API/AbortController)
|
||
|
*
|
||
|
* @returns The response from ChatGPT
|
||
|
*/
|
||
|
sendMessage(text: string, opts?: SendMessageBrowserOptions): Promise<ChatMessage>;
|
||
|
}
|
||
|
|
||
|
export { ChatGPTAPI, ChatGPTAPIOptions, ChatGPTError, ChatGPTUnofficialProxyAPI, ChatMessage, ContentType, ConversationJSONBody, ConversationResponseEvent, CreateChatCompletionStreamResponse, CreateCompletionStreamResponseUsage, FetchFn, GetMessageByIdFunction, Message, MessageActionType, MessageContent, MessageMetadata, Prompt, PromptContent, Role, SendMessageBrowserOptions, SendMessageOptions, UpsertMessageFunction, openai };
|