@markprompt/core
Advanced tools
Comparing version 0.43.0 to 0.43.1
@@ -1,6 +0,6 @@ | ||
import type { Chat, SubmitChatOptions, ChatCompletionMessage, ChatCompletionMessageParam, ChatCompletionMetadata } from './types.js'; | ||
import type { SubmitChatOptions, ChatCompletionChunk, ChatCompletionMessage, ChatCompletionMessageParam, ChatCompletionMetadata } from './types.js'; | ||
import type { BaseOptions } from '../types.js'; | ||
export * from './types.js'; | ||
export * from './utils.js'; | ||
export type SubmitChatYield = Chat.Completions.ChatCompletionChunk.Choice.Delta & ChatCompletionMetadata; | ||
export type SubmitChatYield = ChatCompletionChunk.Choice.Delta & ChatCompletionMetadata; | ||
export type SubmitChatReturn = ChatCompletionMessage & ChatCompletionMetadata; | ||
@@ -7,0 +7,0 @@ export declare const DEFAULT_SUBMIT_CHAT_OPTIONS: { |
@@ -196,3 +196,3 @@ import defaults from 'defaults'; | ||
We wrap the generators with a type ('events' | 'chat') so that we can distingish between them in Promise.race | ||
However, we "lose" the value of whichever generator finishes second in Promise.race | ||
@@ -199,0 +199,0 @@ So, we keep the promise of that each generator.next() returns, and only re-run it |
@@ -1,5 +0,5 @@ | ||
import type { ChatCompletionMessageToolCall, ChatCompletionTool, ChatCompletionToolChoiceOption } from 'openai/resources/chat/index.js'; | ||
import type { ChatCompletionMessageToolCall, ChatCompletionTool, ChatCompletionToolChoiceOption } from 'openai/resources/chat/completions'; | ||
import type { FileSectionReference } from '../types.js'; | ||
import type { ArrayToUnion } from '../utils.js'; | ||
export type { Chat, ChatCompletion, ChatCompletionAssistantMessageParam, ChatCompletionChunk, ChatCompletionContentPart, ChatCompletionFunctionMessageParam, ChatCompletionMessage, ChatCompletionMessageParam, ChatCompletionMessageToolCall, ChatCompletionSystemMessageParam, ChatCompletionTool, ChatCompletionToolChoiceOption, ChatCompletionToolMessageParam, ChatCompletionUserMessageParam, } from 'openai/resources/chat/index.js'; | ||
export type { ChatCompletion, ChatCompletionAssistantMessageParam, ChatCompletionChunk, ChatCompletionContentPart, ChatCompletionContentPartRefusal, ChatCompletionFunctionMessageParam, ChatCompletionMessage, ChatCompletionMessageParam, ChatCompletionMessageToolCall, ChatCompletionSystemMessageParam, ChatCompletionTool, ChatCompletionToolChoiceOption, ChatCompletionToolMessageParam, ChatCompletionUserMessageParam, } from 'openai/resources/chat/completions'; | ||
export declare const CHAT_COMPLETIONS_MODELS: readonly ["claude-3-5-sonnet-20240620", "gpt-3.5-turbo", "gpt-4-1106-preview", "gpt-4-32k", "gpt-4-turbo-preview", "gpt-4", "gpt-4o", "gpt-4o-mini", "meta-llama-3-70b-instruct", "meta-llama-3-8b-instruct", "mixtral-8x7b-instruct-v0.1"]; | ||
@@ -6,0 +6,0 @@ export type ChatCompletionsModel = ArrayToUnion<typeof CHAT_COMPLETIONS_MODELS>; |
@@ -1,2 +0,2 @@ | ||
import type { ChatCompletionContentPart, ChatCompletionContentPartRefusal } from 'openai/resources/index.mjs'; | ||
import type { ChatCompletionMessageParam } from 'openai/resources/chat/completions'; | ||
import type { FileSectionReference } from './types.js'; | ||
@@ -10,4 +10,4 @@ export type RequiredKeys<T, K extends keyof T> = Required<Pick<T, K>> & Omit<T, K>; | ||
export declare function getMessageTextContent(m?: { | ||
content?: null | string | (ChatCompletionContentPart | ChatCompletionContentPartRefusal)[] | readonly (ChatCompletionContentPart | ChatCompletionContentPartRefusal)[]; | ||
content?: null | string | ChatCompletionMessageParam['content']; | ||
}): string | undefined; | ||
//# sourceMappingURL=utils.d.ts.map |
{ | ||
"name": "@markprompt/core", | ||
"version": "0.43.0", | ||
"version": "0.43.1", | ||
"repository": { | ||
@@ -5,0 +5,0 @@ "type": "git", |
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is not supported yet
82388