@promptbook/azure-openai
Advanced tools
Comparing version 0.69.0-15 to 0.69.0-16
@@ -9,4 +9,4 @@ import { OpenAIClient, AzureKeyCredential } from '@azure/openai'; | ||
*/ | ||
var PROMPTBOOK_VERSION = '0.69.0-14'; | ||
// TODO: !!!! List here all the versions and annotate + put into script | ||
var PROMPTBOOK_VERSION = '0.69.0-15'; | ||
// TODO:[main] !!!! List here all the versions and annotate + put into script | ||
@@ -696,3 +696,3 @@ /*! ***************************************************************************** | ||
* TODO: [๐ง ][๐ฃ] More elegant way to tracking than passing `name` | ||
* TODO: [๐ง ] !!! In-memory cache of same values to prevent multiple checks | ||
* TODO: [๐ง ][main] !!! In-memory cache of same values to prevent multiple checks | ||
* Note: [๐ ] This is how `checkSerializableAsJson` + `isSerializableAsJson` together can just retun true/false or rich error message | ||
@@ -1196,2 +1196,3 @@ */ | ||
}, | ||
//TODO:[main] !!!!!! Add gpt-4o-mini-2024-07-18 and all others to be up to date | ||
}, | ||
@@ -1213,2 +1214,47 @@ /**/ | ||
modelVariant: 'CHAT', | ||
modelTitle: 'o1-preview', | ||
modelName: 'o1-preview', | ||
pricing: { | ||
prompt: computeUsage("$15.00 / 1M tokens"), | ||
output: computeUsage("$60.00 / 1M tokens"), | ||
}, | ||
}, | ||
/**/ | ||
/**/ | ||
{ | ||
modelVariant: 'CHAT', | ||
modelTitle: 'o1-preview-2024-09-12', | ||
modelName: 'o1-preview-2024-09-12', | ||
// <- TODO:[main] !!!!!! Some better system to organize theese date suffixes and versions | ||
pricing: { | ||
prompt: computeUsage("$15.00 / 1M tokens"), | ||
output: computeUsage("$60.00 / 1M tokens"), | ||
}, | ||
}, | ||
/**/ | ||
/**/ | ||
{ | ||
modelVariant: 'CHAT', | ||
modelTitle: 'o1-mini', | ||
modelName: 'o1-mini', | ||
pricing: { | ||
prompt: computeUsage("$3.00 / 1M tokens"), | ||
output: computeUsage("$12.00 / 1M tokens"), | ||
}, | ||
}, | ||
/**/ | ||
/**/ | ||
{ | ||
modelVariant: 'CHAT', | ||
modelTitle: 'o1-mini-2024-09-12', | ||
modelName: 'o1-mini-2024-09-12', | ||
pricing: { | ||
prompt: computeUsage("$3.00 / 1M tokens"), | ||
output: computeUsage("$12.00 / 1M tokens"), | ||
}, | ||
}, | ||
/**/ | ||
/**/ | ||
{ | ||
modelVariant: 'CHAT', | ||
modelTitle: 'gpt-3.5-turbo-16k-0613', | ||
@@ -1300,3 +1346,3 @@ modelName: 'gpt-3.5-turbo-16k-0613', | ||
return __generator(this, function (_a) { | ||
// TODO: !!! Do here some filtering which models are really available as deployment | ||
// TODO:[main] !!! Do here some filtering which models are really available as deployment | ||
// @see https://management.azure.com/subscriptions/subscriptionId/resourceGroups/resourceGroupName/providers/Microsoft.CognitiveServices/accounts/accountName/deployments?api-version=2023-05-01 | ||
@@ -1303,0 +1349,0 @@ return [2 /*return*/, OPENAI_MODELS.map(function (_a) { |
declare const _default: ({ | ||
title: string; | ||
pipelineUrl: string; | ||
promptbookVersion: string; | ||
parameters: { | ||
@@ -27,3 +26,2 @@ name: string; | ||
pipelineUrl: string; | ||
promptbookVersion: string; | ||
parameters: { | ||
@@ -57,3 +55,2 @@ name: string; | ||
pipelineUrl: string; | ||
promptbookVersion: string; | ||
parameters: { | ||
@@ -60,0 +57,0 @@ name: string; |
@@ -9,3 +9,3 @@ import type { Command as Program } from 'commander'; | ||
/** | ||
* TODO: [๐ฅ] !!! Allow `ptbk make` without configuring any llm tools | ||
* TODO: [๐ฅ][main] !!! Allow `ptbk make` without configuring any llm tools | ||
* TODO: Maybe remove this command - "about" command should be enough? | ||
@@ -12,0 +12,0 @@ * TODO: [0] DRY Javascript and typescript - Maybe make ONLY typescript and for javascript just remove types |
@@ -28,3 +28,3 @@ import type { string_url } from '../../types/typeAliases'; | ||
/** | ||
* TODO: !!!! [๐ง ] Library precompilation and do not mix markdown and json promptbooks | ||
* TODO:[main] !!!! [๐ง ] Library precompilation and do not mix markdown and json promptbooks | ||
*/ |
@@ -72,3 +72,3 @@ import type { CsvSettings } from './formats/csv/CsvSettings'; | ||
* @@@ | ||
* TODO: [๐] !!! Use | ||
* TODO: [๐][main] !!! Use | ||
* | ||
@@ -80,3 +80,3 @@ * @public exported from `@promptbook/core` | ||
* @@@ | ||
* TODO: [๐] !!! Use | ||
* TODO: [๐][main] !!! Use | ||
* | ||
@@ -83,0 +83,0 @@ * @public exported from `@promptbook/core` |
@@ -21,3 +21,3 @@ import type { PipelineJson } from '../types/PipelineJson/PipelineJson'; | ||
/** | ||
* TODO: !!!! Warn if used only sync version | ||
* TODO:[main] !!!! Warn if used only sync version | ||
* TODO: [๐] Report here line/column of error | ||
@@ -24,0 +24,0 @@ * TODO: Use spaceTrim more effectively |
@@ -35,9 +35,9 @@ import type { PipelineJson } from '../../types/PipelineJson/PipelineJson'; | ||
/** | ||
* TODO: [๐ฃ] !!!! Validate that all samples match expectations | ||
* TODO: [๐ฃ][๐] !!!! Validate that knowledge is valid (non-void) | ||
* TODO: [๐ฃ] !!!! Validate that persona can be used only with CHAT variant | ||
* TODO: [๐ฃ] !!!! Validate that parameter with reserved name not used RESERVED_PARAMETER_NAMES | ||
* TODO: [๐ฃ] !!!! Validate that reserved parameter is not used as joker | ||
* TODO: [๐ฃ][main] !!!! Validate that all samples match expectations | ||
* TODO: [๐ฃ][๐][main] !!!! Validate that knowledge is valid (non-void) | ||
* TODO: [๐ฃ][main] !!!! Validate that persona can be used only with CHAT variant | ||
* TODO: [๐ฃ][main] !!!! Validate that parameter with reserved name not used RESERVED_PARAMETER_NAMES | ||
* TODO: [๐ฃ][main] !!!! Validate that reserved parameter is not used as joker | ||
* TODO: [๐ง ] Validation not only logic itself but imports around - files and websites and rerefenced pipelines exists | ||
* TODO: [๐ ] Actions, instruments (and maybe knowledge) => Functions and tools | ||
*/ |
@@ -11,5 +11,5 @@ import type { PrepareOptions } from '../../../prepare/PrepareOptions'; | ||
/** | ||
* TODO: [๐][๐ผ] !!! Export via `@promptbook/markdown` | ||
* TODO: [๐][๐ผ][main] !!! Export via `@promptbook/markdown` | ||
* TODO: [๐ช] Do it in parallel 11:11 | ||
* Note: No need to aggregate usage here, it is done by intercepting the llmTools | ||
*/ |
@@ -11,3 +11,3 @@ import type { PrepareOptions } from '../../../prepare/PrepareOptions'; | ||
/** | ||
* TODO: [๐][๐ผ] !!! Export via `@promptbook/pdf` | ||
* TODO: [๐][๐ผ][main] !!! Export via `@promptbook/pdf` | ||
* TODO: [๐งบ] In future, content can be alse File or Blob BUT for now for wider compatibility its only base64 | ||
@@ -14,0 +14,0 @@ * @see https://stackoverflow.com/questions/14653349/node-js-cant-create-blobs |
@@ -16,3 +16,3 @@ import type { PromptResult } from '../../../../execution/PromptResult'; | ||
*/ | ||
promptbookVersion: string_promptbook_version; | ||
promptbookVersion?: string_promptbook_version; | ||
/** | ||
@@ -19,0 +19,0 @@ * @@@ |
@@ -19,3 +19,3 @@ import type { AvailableModel } from '../../execution/AvailableModel'; | ||
* Note: [๐ค] Add models of new variant | ||
* TODO: [๐ง ] !!! Add embedding models OR Anthropic has only chat+completion models? | ||
* TODO: [๐ง ][main] !!! Add embedding models OR Anthropic has only chat+completion models? | ||
* TODO: [๐ง ] Some mechanism to propagate unsureness | ||
@@ -22,0 +22,0 @@ * TODO: [๐ง ][๐ฎโโ๏ธ] Put here more info like description, isVision, trainingDateCutoff, languages, strengths ( Top-level performance, intelligence, fluency, and understanding), contextWindow,... |
@@ -14,4 +14,4 @@ import { RemoteLlmExecutionTools } from '../remote/RemoteLlmExecutionTools'; | ||
/** | ||
* TODO: [๐ง ] !!!! Make anonymous this with all LLM providers | ||
* TODO: [๐ง ][๐งฑ] !!!! Maybe change all `new AnthropicClaudeExecutionTools` -> `createAnthropicClaudeExecutionTools` in manual | ||
* TODO: [๐ง ][main] !!!! Make anonymous this with all LLM providers | ||
* TODO: [๐ง ][๐งฑ][main] !!!! Maybe change all `new AnthropicClaudeExecutionTools` -> `createAnthropicClaudeExecutionTools` in manual | ||
* TODO: [๐ง ] Maybe auto-detect usage in browser and determine default value of `isProxied` | ||
@@ -18,0 +18,0 @@ * TODO: [๐ฆบ] Is there some way how to put `packageName` and `className` on top and function definition on bottom? |
#!/usr/bin/env ts-node | ||
export {}; | ||
/** | ||
* TODO: !!! Playground with WebGPT / Promptbook.studio anonymous server | ||
* TODO: !!! Test here that `systemMessage`, `temperature` and `seed` are working correctly | ||
* TODO:[main] !!! Playground with WebGPT / Promptbook.studio anonymous server | ||
* TODO:[main] !!! Test here that `systemMessage`, `temperature` and `seed` are working correctly | ||
*/ |
#!/usr/bin/env ts-node | ||
export {}; | ||
/** | ||
* TODO: !!! Test here that `systemMessage`, `temperature` and `seed` are working correctly | ||
* TODO:[main] !!! Test here that `systemMessage`, `temperature` and `seed` are working correctly | ||
*/ |
@@ -12,3 +12,3 @@ import type { PrepareOptions } from '../prepare/PrepareOptions'; | ||
/** | ||
* TODO: [๐] !!!!! If the persona was prepared with different version or different set of models, prepare it once again | ||
* TODO: [๐][main] !!!!! If the persona was prepared with different version or different set of models, prepare it once again | ||
* TODO: [๐ข] !! Check validity of `modelName` in pipeline | ||
@@ -15,0 +15,0 @@ * TODO: [๐ข] !! Check validity of `systemMessage` in pipeline |
@@ -9,3 +9,3 @@ import type { PipelineJson } from '../types/PipelineJson/PipelineJson'; | ||
/** | ||
* TODO: [๐] !!!!! If the pipeline was prepared with different version or different set of models, prepare it once again | ||
* TODO: [๐][main] !!!!! If the pipeline was prepared with different version or different set of models, prepare it once again | ||
* TODO: [๐ ] Maybe base this on `makeValidator` | ||
@@ -12,0 +12,0 @@ * TODO: [๐ง] Pipeline can be partially prepared, this should return true ONLY if fully prepared |
@@ -26,3 +26,3 @@ import type { PipelineJson } from '../types/PipelineJson/PipelineJson'; | ||
* TODO: [๐ง ] What is better name `prepareTemplate` or `prepareTemplateAndParameters` | ||
* TODO: [โจ] !!! Prepare index the samples and maybe templates | ||
* TODO: [โจ][main] !!! Prepare index the samples and maybe templates | ||
* TODO: Write tests for `preparePipeline` | ||
@@ -29,0 +29,0 @@ * TODO: [๐] Leverage the batch API and build queues @see https://platform.openai.com/docs/guides/batch |
@@ -45,3 +45,3 @@ import type { ModelRequirements } from '../ModelRequirements'; | ||
*/ | ||
readonly promptbookVersion: string_semantic_version; | ||
readonly promptbookVersion?: string_semantic_version; | ||
/** | ||
@@ -48,0 +48,0 @@ * Description of the promptbook |
@@ -595,5 +595,5 @@ import type { TupleToUnion } from 'type-fest'; | ||
/**. | ||
* TODO: !!! Change "For example" to @example | ||
* TODO:[main] !!! Change "For example" to @example | ||
* TODO: !! Change to branded types | ||
* TODO: Delete type aliases that are not exported or used internally | ||
*/ |
@@ -25,4 +25,4 @@ import type { string_name } from '../../types/typeAliases'; | ||
* TODO: [๐ง ][๐ฃ] More elegant way to tracking than passing `name` | ||
* TODO: [๐ง ] !!! In-memory cache of same values to prevent multiple checks | ||
* TODO: [๐ง ][main] !!! In-memory cache of same values to prevent multiple checks | ||
* Note: [๐ ] This is how `checkSerializableAsJson` + `isSerializableAsJson` together can just retun true/false or rich error message | ||
*/ |
@@ -22,4 +22,4 @@ /** | ||
/** | ||
* TODO: [๐ง ] !!! In-memory cache of same values to prevent multiple checks | ||
* TODO: [๐ง ][main] !!! In-memory cache of same values to prevent multiple checks | ||
* TODO: [๐ง ][๐บ] Can be done this on type-level? | ||
*/ |
{ | ||
"name": "@promptbook/azure-openai", | ||
"version": "0.69.0-15", | ||
"version": "0.69.0-16", | ||
"description": "Supercharge your use of large language models", | ||
@@ -50,3 +50,3 @@ "private": false, | ||
"peerDependencies": { | ||
"@promptbook/core": "0.69.0-15" | ||
"@promptbook/core": "0.69.0-16" | ||
}, | ||
@@ -53,0 +53,0 @@ "dependencies": { |
@@ -165,2 +165,6 @@ <!-- โ ๏ธ WARNING: This code has been generated so that any manual changes will be overwritten --> | ||
# โจ New Features | ||
- โจ **Support [OpenAI o1 model](https://openai.com/o1/)** | ||
## ๐ค The Promptbook Whitepaper | ||
@@ -167,0 +171,0 @@ |
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is too big to display
Sorry, the diff of this file is not supported yet
License Policy Violation
LicenseThis package is not allowed per your license policy. Review the package's license to ensure compliance.
Found 1 instance in 1 package
License Policy Violation
LicenseThis package is not allowed per your license policy. Review the package's license to ensure compliance.
Found 1 instance in 1 package
644389
12253
563
504