ollama-ai-provider
Advanced tools
Comparing version 0.12.1 to 0.13.0
@@ -1,4 +0,4 @@ | ||
import { LanguageModelV1, EmbeddingModelV1 } from '@ai-sdk/provider'; | ||
import { LanguageModelV1, ProviderV1, EmbeddingModelV1 } from '@ai-sdk/provider'; | ||
type OllamaChatModelId = 'codegemma' | 'codegemma:2b' | 'codegemma:7b' | 'codellama' | 'codellama:7b' | 'codellama:13b' | 'codellama:34b' | 'codellama:70b' | 'codellama:code' | 'codellama:python' | 'command-r' | 'command-r:35b' | 'command-r-plus' | 'command-r-plus:104b' | 'deepseek-coder-v2' | 'deepseek-coder-v2:16b' | 'deepseek-coder-v2:236b' | 'falcon2' | 'falcon2:11b' | 'gemma' | 'gemma:2b' | 'gemma:7b' | 'gemma2' | 'gemma2:2b' | 'gemma2:9b' | 'gemma2:27b' | 'llama2' | 'llama2:7b' | 'llama2:13b' | 'llama2:70b' | 'llama3' | 'llama3:8b' | 'llama3:70b' | 'llama3-chatqa' | 'llama3-chatqa:8b' | 'llama3-chatqa:70b' | 'llama3-gradient' | 'llama3-gradient:8b' | 'llama3-gradient:70b' | 'llama3.1' | 'llama3.1:8b' | 'llama3.1:70b' | 'llama3.1:405b' | 'llava' | 'llava:7b' | 'llava:13b' | 'llava:34b' | 'llava-llama3' | 'llava-llama3:8b' | 'llava-phi3' | 'llava-phi3:3.8b' | 'mistral' | 'mistral:7b' | 'mistral-large' | 'mistral-large:123b' | 'mistral-nemo' | 'mistral-nemo:12b' | 'mixtral' | 'mixtral:8x7b' | 'mixtral:8x22b' | 'moondream' | 'moondream:1.8b' | 'openhermes' | 'openhermes:v2.5' | 'phi3' | 'phi3:3.8b' | 'phi3:14b' | 'phi3.5' | 'phi3.5:3.8b' | 'qwen' | 'qwen:7b' | 'qwen:14b' | 'qwen:32b' | 'qwen:72b' | 'qwen:110b' | 'qwen2' | 'qwen2:0.5b' | 'qwen2:1.5b' | 'qwen2:7b' | 'qwen2:72b' | 'smollm' | 'smollm:135m' | 'smollm:360m' | 'smollm:1.7b' | (string & NonNullable<unknown>); | ||
type OllamaChatModelId = 'codegemma' | 'codegemma:2b' | 'codegemma:7b' | 'codellama' | 'codellama:7b' | 'codellama:13b' | 'codellama:34b' | 'codellama:70b' | 'codellama:code' | 'codellama:python' | 'command-r' | 'command-r:35b' | 'command-r-plus' | 'command-r-plus:104b' | 'deepseek-coder-v2' | 'deepseek-coder-v2:16b' | 'deepseek-coder-v2:236b' | 'falcon2' | 'falcon2:11b' | 'firefunction-v2' | 'firefunction-v2:70b' | 'gemma' | 'gemma:2b' | 'gemma:7b' | 'gemma2' | 'gemma2:2b' | 'gemma2:9b' | 'gemma2:27b' | 'llama2' | 'llama2:7b' | 'llama2:13b' | 'llama2:70b' | 'llama3' | 'llama3:8b' | 'llama3:70b' | 'llama3-chatqa' | 'llama3-chatqa:8b' | 'llama3-chatqa:70b' | 'llama3-gradient' | 'llama3-gradient:8b' | 'llama3-gradient:70b' | 'llama3.1' | 'llama3.1:8b' | 'llama3.1:70b' | 'llama3.1:405b' | 'llava' | 'llava:7b' | 'llava:13b' | 'llava:34b' | 'llava-llama3' | 'llava-llama3:8b' | 'llava-phi3' | 'llava-phi3:3.8b' | 'mistral' | 'mistral:7b' | 'mistral-large' | 'mistral-large:123b' | 'mistral-nemo' | 'mistral-nemo:12b' | 'mixtral' | 'mixtral:8x7b' | 'mixtral:8x22b' | 'moondream' | 'moondream:1.8b' | 'openhermes' | 'openhermes:v2.5' | 'phi3' | 'phi3:3.8b' | 'phi3:14b' | 'phi3.5' | 'phi3.5:3.8b' | 'qwen' | 'qwen:7b' | 'qwen:14b' | 'qwen:32b' | 'qwen:72b' | 'qwen:110b' | 'qwen2' | 'qwen2:0.5b' | 'qwen2:1.5b' | 'qwen2:7b' | 'qwen2:72b' | 'smollm' | 'smollm:135m' | 'smollm:360m' | 'smollm:1.7b' | (string & NonNullable<unknown>); | ||
interface OllamaChatSettings { | ||
@@ -84,27 +84,16 @@ /** | ||
type OllamaEmbeddingConfig = { | ||
baseURL: string; | ||
fetch?: typeof fetch; | ||
headers: () => Record<string, string | undefined>; | ||
provider: string; | ||
}; | ||
declare class OllamaEmbeddingModel implements EmbeddingModelV1<string> { | ||
readonly specificationVersion = "v1"; | ||
readonly modelId: OllamaEmbeddingModelId; | ||
private readonly config; | ||
private readonly settings; | ||
get provider(): string; | ||
get maxEmbeddingsPerCall(): number; | ||
get supportsParallelCalls(): boolean; | ||
constructor(modelId: OllamaEmbeddingModelId, settings: OllamaEmbeddingSettings, config: OllamaEmbeddingConfig); | ||
doEmbed({ abortSignal, values, }: Parameters<EmbeddingModelV1<string>['doEmbed']>[0]): Promise<Awaited<ReturnType<EmbeddingModelV1<string>['doEmbed']>>>; | ||
interface OllamaProvider extends ProviderV1 { | ||
(modelId: OllamaChatModelId, settings?: OllamaChatSettings): LanguageModelV1; | ||
chat(modelId: OllamaChatModelId, settings?: OllamaChatSettings): LanguageModelV1; | ||
/** | ||
* @deprecated Use `textEmbeddingModel` instead. | ||
*/ | ||
embedding(modelId: OllamaEmbeddingModelId, settings?: OllamaEmbeddingSettings): EmbeddingModelV1<string>; | ||
languageModel(modelId: OllamaChatModelId, settings?: OllamaChatSettings): LanguageModelV1; | ||
/** | ||
* @deprecated Use `textEmbeddingModel()` instead. | ||
*/ | ||
textEmbedding(modelId: OllamaEmbeddingModelId, settings?: OllamaEmbeddingSettings): EmbeddingModelV1<string>; | ||
textEmbeddingModel(modelId: OllamaEmbeddingModelId, settings?: OllamaEmbeddingSettings): EmbeddingModelV1<string>; | ||
} | ||
interface OllamaProvider { | ||
(modelId: OllamaChatModelId, settings?: OllamaChatSettings): OllamaChatLanguageModel; | ||
chat(modelId: OllamaChatModelId, settings?: OllamaChatSettings): OllamaChatLanguageModel; | ||
embedding(modelId: OllamaEmbeddingModelId, settings?: OllamaEmbeddingSettings): OllamaEmbeddingModel; | ||
languageModel(modelId: OllamaChatModelId, settings?: OllamaChatSettings): OllamaChatLanguageModel; | ||
textEmbedding(modelId: OllamaEmbeddingModelId, settings?: OllamaEmbeddingSettings): OllamaEmbeddingModel; | ||
} | ||
interface OllamaProviderSettings { | ||
@@ -111,0 +100,0 @@ /** |
@@ -706,2 +706,3 @@ "use strict"; | ||
provider.textEmbedding = createEmbeddingModel; | ||
provider.textEmbeddingModel = createEmbeddingModel; | ||
return provider; | ||
@@ -708,0 +709,0 @@ } |
{ | ||
"name": "ollama-ai-provider", | ||
"version": "0.12.1", | ||
"version": "0.13.0", | ||
"description": "Vercel AI Provider for running LLMs locally using Ollama", | ||
@@ -18,4 +18,4 @@ "main": "./dist/index.js", | ||
"dependencies": { | ||
"@ai-sdk/provider": "0.0.21", | ||
"@ai-sdk/provider-utils": "1.0.16", | ||
"@ai-sdk/provider": "0.0.22", | ||
"@ai-sdk/provider-utils": "1.0.17", | ||
"partial-json": "0.1.7" | ||
@@ -25,3 +25,3 @@ }, | ||
"@edge-runtime/vm": "^3.2.0", | ||
"@types/node": "^18.19.43", | ||
"@types/node": "^18.19.46", | ||
"tsup": "^8.2.4", | ||
@@ -28,0 +28,0 @@ "typescript": "5.5.4", |
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is not supported yet
141356
1509
+ Added@ai-sdk/provider@0.0.22(transitive)
+ Added@ai-sdk/provider-utils@1.0.17(transitive)
- Removed@ai-sdk/provider@0.0.21(transitive)
- Removed@ai-sdk/provider-utils@1.0.16(transitive)
Updated@ai-sdk/provider@0.0.22