@ai-sdk/solid
Advanced tools
Comparing version 0.0.54 to 1.0.0-canary.3
@@ -32,3 +32,3 @@ "use strict"; | ||
var import_store = require("solid-js/store"); | ||
var getStreamedResponse = async (api, chatRequest, mutate, setStreamData, streamData, extraMetadata, messagesRef, abortController, generateId, streamProtocol, onFinish, onResponse, onToolCall, sendExtraMessageFields, fetch, keepLastMessageOnError) => { | ||
var processStreamedResponse = async (api, chatRequest, mutate, setStreamData, streamData, extraMetadata, messagesRef, abortController, generateId, streamProtocol = "data", onFinish, onResponse, onToolCall, sendExtraMessageFields, fetch, keepLastMessageOnError) => { | ||
var _a; | ||
@@ -39,6 +39,5 @@ const previousMessages = messagesRef; | ||
const constructedMessagesPayload = sendExtraMessageFields ? chatRequest.messages : chatRequest.messages.map( | ||
({ role, content, name, data, annotations, toolInvocations }) => ({ | ||
({ role, content, data, annotations, toolInvocations }) => ({ | ||
role, | ||
content, | ||
...name !== void 0 && { name }, | ||
...data !== void 0 && { data }, | ||
@@ -134,3 +133,3 @@ ...annotations !== void 0 && { annotations }, | ||
const triggerRequest = async (chatRequest) => { | ||
var _a2, _b2, _c2, _d2, _e2, _f2, _g2, _h2, _i, _j; | ||
var _a2, _b2, _c2, _d2, _e2, _f2, _g2, _h2, _i, _j, _k, _l, _m, _n, _o, _p, _q, _r, _s, _t; | ||
const messageCount = messagesRef.length; | ||
@@ -141,31 +140,20 @@ try { | ||
abortController = new AbortController(); | ||
await (0, import_ui_utils.processChatStream)({ | ||
getStreamedResponse: () => { | ||
var _a3, _b3, _c3, _d3, _e3, _f3, _g3, _h3, _i2, _j2, _k, _l, _m, _n, _o, _p, _q, _r; | ||
return getStreamedResponse( | ||
api(), | ||
chatRequest, | ||
mutate, | ||
setStreamData, | ||
streamData, | ||
extraMetadata, | ||
messagesRef, | ||
abortController, | ||
generateId(), | ||
// streamMode is deprecated, use streamProtocol instead: | ||
((_e3 = (_b3 = (_a3 = useChatOptions()).streamProtocol) == null ? void 0 : _b3.call(_a3)) != null ? _e3 : ((_d3 = (_c3 = useChatOptions()).streamMode) == null ? void 0 : _d3.call(_c3)) === "text") ? "text" : void 0, | ||
(_g3 = (_f3 = useChatOptions()).onFinish) == null ? void 0 : _g3.call(_f3), | ||
(_i2 = (_h3 = useChatOptions()).onResponse) == null ? void 0 : _i2.call(_h3), | ||
(_k = (_j2 = useChatOptions()).onToolCall) == null ? void 0 : _k.call(_j2), | ||
(_m = (_l = useChatOptions()).sendExtraMessageFields) == null ? void 0 : _m.call(_l), | ||
(_o = (_n = useChatOptions()).fetch) == null ? void 0 : _o.call(_n), | ||
(_r = (_q = (_p = useChatOptions()).keepLastMessageOnError) == null ? void 0 : _q.call(_p)) != null ? _r : false | ||
); | ||
}, | ||
experimental_onFunctionCall: (_b2 = (_a2 = useChatOptions()).experimental_onFunctionCall) == null ? void 0 : _b2.call(_a2), | ||
updateChatRequest(newChatRequest) { | ||
chatRequest = newChatRequest; | ||
}, | ||
getCurrentMessages: () => messagesRef | ||
}); | ||
await processStreamedResponse( | ||
api(), | ||
chatRequest, | ||
mutate, | ||
setStreamData, | ||
streamData, | ||
extraMetadata, | ||
messagesRef, | ||
abortController, | ||
generateId(), | ||
(_b2 = (_a2 = useChatOptions()).streamProtocol) == null ? void 0 : _b2.call(_a2), | ||
(_d2 = (_c2 = useChatOptions()).onFinish) == null ? void 0 : _d2.call(_c2), | ||
(_f2 = (_e2 = useChatOptions()).onResponse) == null ? void 0 : _f2.call(_e2), | ||
(_h2 = (_g2 = useChatOptions()).onToolCall) == null ? void 0 : _h2.call(_g2), | ||
(_j = (_i = useChatOptions()).sendExtraMessageFields) == null ? void 0 : _j.call(_i), | ||
(_l = (_k = useChatOptions()).fetch) == null ? void 0 : _l.call(_k), | ||
(_o = (_n = (_m = useChatOptions()).keepLastMessageOnError) == null ? void 0 : _n.call(_m)) != null ? _o : false | ||
); | ||
abortController = null; | ||
@@ -177,3 +165,3 @@ } catch (err) { | ||
} | ||
const onError = (_d2 = (_c2 = useChatOptions()).onError) == null ? void 0 : _d2.call(_c2); | ||
const onError = (_q = (_p = useChatOptions()).onError) == null ? void 0 : _q.call(_p); | ||
if (onError && err instanceof Error) { | ||
@@ -186,3 +174,3 @@ onError(err); | ||
} | ||
const maxSteps = (_j = (_f2 = (_e2 = useChatOptions()).maxSteps) == null ? void 0 : _f2.call(_e2)) != null ? _j : ((_i = (_h2 = (_g2 = useChatOptions()).maxToolRoundtrips) == null ? void 0 : _h2.call(_g2)) != null ? _i : 0) + 1; | ||
const maxSteps = (_t = (_s = (_r = useChatOptions()).maxSteps) == null ? void 0 : _s.call(_r)) != null ? _t : 1; | ||
const messages2 = messagesRef; | ||
@@ -201,21 +189,14 @@ const lastMessage = messages2[messages2.length - 1]; | ||
}; | ||
const append = async (message, { options, data, headers, body } = {}) => { | ||
const append = async (message, { data, headers, body } = {}) => { | ||
if (!message.id) { | ||
message.id = generateId()(); | ||
} | ||
const requestOptions = { | ||
headers: headers != null ? headers : options == null ? void 0 : options.headers, | ||
body: body != null ? body : options == null ? void 0 : options.body | ||
}; | ||
const chatRequest = { | ||
return triggerRequest({ | ||
messages: messagesRef.concat(message), | ||
options: requestOptions, | ||
headers: requestOptions.headers, | ||
body: requestOptions.body, | ||
headers, | ||
body, | ||
data | ||
}; | ||
return triggerRequest(chatRequest); | ||
}); | ||
}; | ||
const reload = async ({ | ||
options, | ||
data, | ||
@@ -225,27 +206,12 @@ headers, | ||
} = {}) => { | ||
if (messagesRef.length === 0) | ||
if (messagesRef.length === 0) { | ||
return null; | ||
const requestOptions = { | ||
headers: headers != null ? headers : options == null ? void 0 : options.headers, | ||
body: body != null ? body : options == null ? void 0 : options.body | ||
}; | ||
} | ||
const lastMessage = messagesRef[messagesRef.length - 1]; | ||
if (lastMessage.role === "assistant") { | ||
const chatRequest2 = { | ||
messages: messagesRef.slice(0, -1), | ||
options: requestOptions, | ||
headers: requestOptions.headers, | ||
body: requestOptions.body, | ||
data | ||
}; | ||
return triggerRequest(chatRequest2); | ||
} | ||
const chatRequest = { | ||
messages: messagesRef, | ||
options: requestOptions, | ||
headers: requestOptions.headers, | ||
body: requestOptions.body, | ||
return triggerRequest({ | ||
messages: lastMessage.role === "assistant" ? messagesRef.slice(0, -1) : messagesRef, | ||
headers, | ||
body, | ||
data | ||
}; | ||
return triggerRequest(chatRequest); | ||
}); | ||
}; | ||
@@ -275,3 +241,3 @@ const stop = () => { | ||
const handleSubmit = (event, options = {}, metadata) => { | ||
var _a2, _b2, _c2, _d2, _e2; | ||
var _a2; | ||
(_a2 = event == null ? void 0 : event.preventDefault) == null ? void 0 : _a2.call(event); | ||
@@ -287,7 +253,3 @@ const inputValue = input(); | ||
} | ||
const requestOptions = { | ||
headers: (_c2 = options.headers) != null ? _c2 : (_b2 = options.options) == null ? void 0 : _b2.headers, | ||
body: (_e2 = options.body) != null ? _e2 : (_d2 = options.options) == null ? void 0 : _d2.body | ||
}; | ||
const chatRequest = { | ||
triggerRequest({ | ||
messages: !inputValue && options.allowEmptySubmit ? messagesRef : messagesRef.concat({ | ||
@@ -299,8 +261,6 @@ id: generateId()(), | ||
}), | ||
options: requestOptions, | ||
body: requestOptions.body, | ||
headers: requestOptions.headers, | ||
headers: options.headers, | ||
body: options.body, | ||
data: options.data | ||
}; | ||
triggerRequest(chatRequest); | ||
}); | ||
setInput(""); | ||
@@ -430,3 +390,3 @@ }; | ||
const complete = async (prompt, options) => { | ||
var _a2, _b2, _c2, _d2, _e2, _f2, _g2, _h2, _i2, _j, _k, _l, _m, _n, _o, _p; | ||
var _a2, _b2, _c2, _d2, _e2, _f2, _g2, _h2, _i2, _j, _k, _l, _m; | ||
const existingData = (_a2 = streamData()) != null ? _a2 : []; | ||
@@ -442,4 +402,3 @@ return (0, import_ui_utils2.callCompletionApi)({ | ||
}, | ||
// streamMode is deprecated, use streamProtocol instead: | ||
streamProtocol: ((_h2 = (_e2 = (_d2 = useCompletionOptions()).streamProtocol) == null ? void 0 : _e2.call(_d2)) != null ? _h2 : ((_g2 = (_f2 = useCompletionOptions()).streamMode) == null ? void 0 : _g2.call(_f2)) === "text") ? "text" : void 0, | ||
streamProtocol: (_e2 = (_d2 = useCompletionOptions()).streamProtocol) == null ? void 0 : _e2.call(_d2), | ||
setCompletion: mutate, | ||
@@ -449,9 +408,9 @@ setLoading: setIsLoading, | ||
setAbortController, | ||
onResponse: (_j = (_i2 = useCompletionOptions()).onResponse) == null ? void 0 : _j.call(_i2), | ||
onFinish: (_l = (_k = useCompletionOptions()).onFinish) == null ? void 0 : _l.call(_k), | ||
onError: (_n = (_m = useCompletionOptions()).onError) == null ? void 0 : _n.call(_m), | ||
onResponse: (_g2 = (_f2 = useCompletionOptions()).onResponse) == null ? void 0 : _g2.call(_f2), | ||
onFinish: (_i2 = (_h2 = useCompletionOptions()).onFinish) == null ? void 0 : _i2.call(_h2), | ||
onError: (_k = (_j = useCompletionOptions()).onError) == null ? void 0 : _k.call(_j), | ||
onData: (data) => { | ||
setStreamData([...existingData, ...data != null ? data : []]); | ||
}, | ||
fetch: (_p = (_o = useCompletionOptions()).fetch) == null ? void 0 : _p.call(_o) | ||
fetch: (_m = (_l = useCompletionOptions()).fetch) == null ? void 0 : _m.call(_l) | ||
}); | ||
@@ -458,0 +417,0 @@ }; |
# @ai-sdk/solid | ||
## 1.0.0-canary.3 | ||
### Patch Changes | ||
- Updated dependencies [b1da952] | ||
- @ai-sdk/provider-utils@2.0.0-canary.1 | ||
- @ai-sdk/ui-utils@1.0.0-canary.3 | ||
## 1.0.0-canary.2 | ||
### Major Changes | ||
- e117b54: chore (ui): remove deprecated useChat roundtrip options | ||
- 7814c4b: chore (ui): remove streamMode setting from useChat & useCompletion | ||
### Patch Changes | ||
- Updated dependencies [b469a7e] | ||
- Updated dependencies [7814c4b] | ||
- Updated dependencies [db46ce5] | ||
- @ai-sdk/provider-utils@2.0.0-canary.0 | ||
- @ai-sdk/ui-utils@1.0.0-canary.2 | ||
## 1.0.0-canary.1 | ||
### Major Changes | ||
- 8bf5756: chore: remove legacy function/tool calling | ||
### Patch Changes | ||
- Updated dependencies [8bf5756] | ||
- @ai-sdk/ui-utils@1.0.0-canary.1 | ||
## 1.0.0-canary.0 | ||
### Major Changes | ||
- 84edae5: chore (release): bump ui package versions for 4.0 release | ||
### Patch Changes | ||
- Updated dependencies [7e89ccb] | ||
- @ai-sdk/ui-utils@1.0.0-canary.0 | ||
## 0.0.54 | ||
@@ -4,0 +49,0 @@ |
@@ -58,17 +58,2 @@ import { FetchFunction } from '@ai-sdk/provider-utils'; | ||
/** | ||
Maximum number of automatic roundtrips for tool calls. | ||
An automatic tool call roundtrip is a call to the server with the | ||
tool call results when all tool calls in the last assistant | ||
message have results. | ||
A maximum number is required to prevent infinite loops in the | ||
case of misconfigured tools. | ||
By default, it's set to 0, which will disable the feature. | ||
@deprecated Use `maxSteps` instead (which is `maxToolRoundtrips` + 1). | ||
*/ | ||
maxToolRoundtrips?: number; | ||
/** | ||
Maximum number of sequential LLM calls (steps), e.g. when you use tool calls. Must be at least 1. | ||
@@ -75,0 +60,0 @@ |
@@ -32,3 +32,3 @@ "use strict"; | ||
var import_store = require("solid-js/store"); | ||
var getStreamedResponse = async (api, chatRequest, mutate, setStreamData, streamData, extraMetadata, messagesRef, abortController, generateId, streamProtocol, onFinish, onResponse, onToolCall, sendExtraMessageFields, fetch, keepLastMessageOnError) => { | ||
var processStreamedResponse = async (api, chatRequest, mutate, setStreamData, streamData, extraMetadata, messagesRef, abortController, generateId, streamProtocol = "data", onFinish, onResponse, onToolCall, sendExtraMessageFields, fetch, keepLastMessageOnError) => { | ||
var _a; | ||
@@ -39,6 +39,5 @@ const previousMessages = messagesRef; | ||
const constructedMessagesPayload = sendExtraMessageFields ? chatRequest.messages : chatRequest.messages.map( | ||
({ role, content, name, data, annotations, toolInvocations }) => ({ | ||
({ role, content, data, annotations, toolInvocations }) => ({ | ||
role, | ||
content, | ||
...name !== void 0 && { name }, | ||
...data !== void 0 && { data }, | ||
@@ -134,3 +133,3 @@ ...annotations !== void 0 && { annotations }, | ||
const triggerRequest = async (chatRequest) => { | ||
var _a2, _b2, _c2, _d2, _e2, _f2, _g2, _h2, _i, _j; | ||
var _a2, _b2, _c2, _d2, _e2, _f2, _g2, _h2, _i, _j, _k, _l, _m, _n, _o, _p, _q, _r, _s, _t; | ||
const messageCount = messagesRef.length; | ||
@@ -141,31 +140,20 @@ try { | ||
abortController = new AbortController(); | ||
await (0, import_ui_utils.processChatStream)({ | ||
getStreamedResponse: () => { | ||
var _a3, _b3, _c3, _d3, _e3, _f3, _g3, _h3, _i2, _j2, _k, _l, _m, _n, _o, _p, _q, _r; | ||
return getStreamedResponse( | ||
api(), | ||
chatRequest, | ||
mutate, | ||
setStreamData, | ||
streamData, | ||
extraMetadata, | ||
messagesRef, | ||
abortController, | ||
generateId(), | ||
// streamMode is deprecated, use streamProtocol instead: | ||
((_e3 = (_b3 = (_a3 = useChatOptions()).streamProtocol) == null ? void 0 : _b3.call(_a3)) != null ? _e3 : ((_d3 = (_c3 = useChatOptions()).streamMode) == null ? void 0 : _d3.call(_c3)) === "text") ? "text" : void 0, | ||
(_g3 = (_f3 = useChatOptions()).onFinish) == null ? void 0 : _g3.call(_f3), | ||
(_i2 = (_h3 = useChatOptions()).onResponse) == null ? void 0 : _i2.call(_h3), | ||
(_k = (_j2 = useChatOptions()).onToolCall) == null ? void 0 : _k.call(_j2), | ||
(_m = (_l = useChatOptions()).sendExtraMessageFields) == null ? void 0 : _m.call(_l), | ||
(_o = (_n = useChatOptions()).fetch) == null ? void 0 : _o.call(_n), | ||
(_r = (_q = (_p = useChatOptions()).keepLastMessageOnError) == null ? void 0 : _q.call(_p)) != null ? _r : false | ||
); | ||
}, | ||
experimental_onFunctionCall: (_b2 = (_a2 = useChatOptions()).experimental_onFunctionCall) == null ? void 0 : _b2.call(_a2), | ||
updateChatRequest(newChatRequest) { | ||
chatRequest = newChatRequest; | ||
}, | ||
getCurrentMessages: () => messagesRef | ||
}); | ||
await processStreamedResponse( | ||
api(), | ||
chatRequest, | ||
mutate, | ||
setStreamData, | ||
streamData, | ||
extraMetadata, | ||
messagesRef, | ||
abortController, | ||
generateId(), | ||
(_b2 = (_a2 = useChatOptions()).streamProtocol) == null ? void 0 : _b2.call(_a2), | ||
(_d2 = (_c2 = useChatOptions()).onFinish) == null ? void 0 : _d2.call(_c2), | ||
(_f2 = (_e2 = useChatOptions()).onResponse) == null ? void 0 : _f2.call(_e2), | ||
(_h2 = (_g2 = useChatOptions()).onToolCall) == null ? void 0 : _h2.call(_g2), | ||
(_j = (_i = useChatOptions()).sendExtraMessageFields) == null ? void 0 : _j.call(_i), | ||
(_l = (_k = useChatOptions()).fetch) == null ? void 0 : _l.call(_k), | ||
(_o = (_n = (_m = useChatOptions()).keepLastMessageOnError) == null ? void 0 : _n.call(_m)) != null ? _o : false | ||
); | ||
abortController = null; | ||
@@ -177,3 +165,3 @@ } catch (err) { | ||
} | ||
const onError = (_d2 = (_c2 = useChatOptions()).onError) == null ? void 0 : _d2.call(_c2); | ||
const onError = (_q = (_p = useChatOptions()).onError) == null ? void 0 : _q.call(_p); | ||
if (onError && err instanceof Error) { | ||
@@ -186,3 +174,3 @@ onError(err); | ||
} | ||
const maxSteps = (_j = (_f2 = (_e2 = useChatOptions()).maxSteps) == null ? void 0 : _f2.call(_e2)) != null ? _j : ((_i = (_h2 = (_g2 = useChatOptions()).maxToolRoundtrips) == null ? void 0 : _h2.call(_g2)) != null ? _i : 0) + 1; | ||
const maxSteps = (_t = (_s = (_r = useChatOptions()).maxSteps) == null ? void 0 : _s.call(_r)) != null ? _t : 1; | ||
const messages2 = messagesRef; | ||
@@ -201,21 +189,14 @@ const lastMessage = messages2[messages2.length - 1]; | ||
}; | ||
const append = async (message, { options, data, headers, body } = {}) => { | ||
const append = async (message, { data, headers, body } = {}) => { | ||
if (!message.id) { | ||
message.id = generateId()(); | ||
} | ||
const requestOptions = { | ||
headers: headers != null ? headers : options == null ? void 0 : options.headers, | ||
body: body != null ? body : options == null ? void 0 : options.body | ||
}; | ||
const chatRequest = { | ||
return triggerRequest({ | ||
messages: messagesRef.concat(message), | ||
options: requestOptions, | ||
headers: requestOptions.headers, | ||
body: requestOptions.body, | ||
headers, | ||
body, | ||
data | ||
}; | ||
return triggerRequest(chatRequest); | ||
}); | ||
}; | ||
const reload = async ({ | ||
options, | ||
data, | ||
@@ -225,27 +206,12 @@ headers, | ||
} = {}) => { | ||
if (messagesRef.length === 0) | ||
if (messagesRef.length === 0) { | ||
return null; | ||
const requestOptions = { | ||
headers: headers != null ? headers : options == null ? void 0 : options.headers, | ||
body: body != null ? body : options == null ? void 0 : options.body | ||
}; | ||
} | ||
const lastMessage = messagesRef[messagesRef.length - 1]; | ||
if (lastMessage.role === "assistant") { | ||
const chatRequest2 = { | ||
messages: messagesRef.slice(0, -1), | ||
options: requestOptions, | ||
headers: requestOptions.headers, | ||
body: requestOptions.body, | ||
data | ||
}; | ||
return triggerRequest(chatRequest2); | ||
} | ||
const chatRequest = { | ||
messages: messagesRef, | ||
options: requestOptions, | ||
headers: requestOptions.headers, | ||
body: requestOptions.body, | ||
return triggerRequest({ | ||
messages: lastMessage.role === "assistant" ? messagesRef.slice(0, -1) : messagesRef, | ||
headers, | ||
body, | ||
data | ||
}; | ||
return triggerRequest(chatRequest); | ||
}); | ||
}; | ||
@@ -275,3 +241,3 @@ const stop = () => { | ||
const handleSubmit = (event, options = {}, metadata) => { | ||
var _a2, _b2, _c2, _d2, _e2; | ||
var _a2; | ||
(_a2 = event == null ? void 0 : event.preventDefault) == null ? void 0 : _a2.call(event); | ||
@@ -287,7 +253,3 @@ const inputValue = input(); | ||
} | ||
const requestOptions = { | ||
headers: (_c2 = options.headers) != null ? _c2 : (_b2 = options.options) == null ? void 0 : _b2.headers, | ||
body: (_e2 = options.body) != null ? _e2 : (_d2 = options.options) == null ? void 0 : _d2.body | ||
}; | ||
const chatRequest = { | ||
triggerRequest({ | ||
messages: !inputValue && options.allowEmptySubmit ? messagesRef : messagesRef.concat({ | ||
@@ -299,8 +261,6 @@ id: generateId()(), | ||
}), | ||
options: requestOptions, | ||
body: requestOptions.body, | ||
headers: requestOptions.headers, | ||
headers: options.headers, | ||
body: options.body, | ||
data: options.data | ||
}; | ||
triggerRequest(chatRequest); | ||
}); | ||
setInput(""); | ||
@@ -430,3 +390,3 @@ }; | ||
const complete = async (prompt, options) => { | ||
var _a2, _b2, _c2, _d2, _e2, _f2, _g2, _h2, _i2, _j, _k, _l, _m, _n, _o, _p; | ||
var _a2, _b2, _c2, _d2, _e2, _f2, _g2, _h2, _i2, _j, _k, _l, _m; | ||
const existingData = (_a2 = streamData()) != null ? _a2 : []; | ||
@@ -442,4 +402,3 @@ return (0, import_ui_utils2.callCompletionApi)({ | ||
}, | ||
// streamMode is deprecated, use streamProtocol instead: | ||
streamProtocol: ((_h2 = (_e2 = (_d2 = useCompletionOptions()).streamProtocol) == null ? void 0 : _e2.call(_d2)) != null ? _h2 : ((_g2 = (_f2 = useCompletionOptions()).streamMode) == null ? void 0 : _g2.call(_f2)) === "text") ? "text" : void 0, | ||
streamProtocol: (_e2 = (_d2 = useCompletionOptions()).streamProtocol) == null ? void 0 : _e2.call(_d2), | ||
setCompletion: mutate, | ||
@@ -449,9 +408,9 @@ setLoading: setIsLoading, | ||
setAbortController, | ||
onResponse: (_j = (_i2 = useCompletionOptions()).onResponse) == null ? void 0 : _j.call(_i2), | ||
onFinish: (_l = (_k = useCompletionOptions()).onFinish) == null ? void 0 : _l.call(_k), | ||
onError: (_n = (_m = useCompletionOptions()).onError) == null ? void 0 : _n.call(_m), | ||
onResponse: (_g2 = (_f2 = useCompletionOptions()).onResponse) == null ? void 0 : _g2.call(_f2), | ||
onFinish: (_i2 = (_h2 = useCompletionOptions()).onFinish) == null ? void 0 : _i2.call(_h2), | ||
onError: (_k = (_j = useCompletionOptions()).onError) == null ? void 0 : _k.call(_j), | ||
onData: (data) => { | ||
setStreamData([...existingData, ...data != null ? data : []]); | ||
}, | ||
fetch: (_p = (_o = useCompletionOptions()).fetch) == null ? void 0 : _p.call(_o) | ||
fetch: (_m = (_l = useCompletionOptions()).fetch) == null ? void 0 : _m.call(_l) | ||
}); | ||
@@ -458,0 +417,0 @@ }; |
{ | ||
"name": "@ai-sdk/solid", | ||
"version": "0.0.54", | ||
"version": "1.0.0-canary.3", | ||
"license": "Apache-2.0", | ||
@@ -22,4 +22,4 @@ "sideEffects": false, | ||
"dependencies": { | ||
"@ai-sdk/provider-utils": "1.0.22", | ||
"@ai-sdk/ui-utils": "0.0.50" | ||
"@ai-sdk/provider-utils": "2.0.0-canary.1", | ||
"@ai-sdk/ui-utils": "1.0.0-canary.3" | ||
}, | ||
@@ -26,0 +26,0 @@ "devDependencies": { |
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is not supported yet
License Policy Violation
LicenseThis package is not allowed per your license policy. Review the package's license to ensure compliance.
Found 1 instance in 1 package
License Policy Violation
LicenseThis package is not allowed per your license policy. Review the package's license to ensure compliance.
Found 1 instance in 1 package
128753
1452
+ Added@ai-sdk/provider@1.0.0-canary.0(transitive)
+ Added@ai-sdk/provider-utils@2.0.0-canary.1(transitive)
+ Added@ai-sdk/ui-utils@1.0.0-canary.3(transitive)
- Removed@ai-sdk/provider@0.0.26(transitive)
- Removed@ai-sdk/provider-utils@1.0.22(transitive)
- Removed@ai-sdk/ui-utils@0.0.50(transitive)