@baseai/core
Advanced tools
Comparing version 0.9.15-snapshot.11 to 0.9.15-snapshot.12
@@ -20,5 +20,5 @@ "use strict"; | ||
// index.ts | ||
var core_exports = {}; | ||
__export(core_exports, { | ||
// src/index.ts | ||
var src_exports = {}; | ||
__export(src_exports, { | ||
Pipe: () => Pipe, | ||
@@ -38,5 +38,5 @@ fromReadableStream: () => fromReadableStream, | ||
}); | ||
module.exports = __toCommonJS(core_exports); | ||
module.exports = __toCommonJS(src_exports); | ||
// helpers/stream.ts | ||
// src/helpers/stream.ts | ||
var import_ChatCompletionStream = require("openai/lib/ChatCompletionStream"); | ||
@@ -54,6 +54,6 @@ var fromReadableStream = (readableStream) => { | ||
// common/request.ts | ||
// src/common/request.ts | ||
var import_streaming = require("openai/streaming"); | ||
// common/errors.ts | ||
// src/common/errors.ts | ||
var APIError = class _APIError extends Error { | ||
@@ -172,3 +172,3 @@ constructor(status, error, message, headers) { | ||
// common/request.ts | ||
// src/common/request.ts | ||
var Request = class { | ||
@@ -308,3 +308,3 @@ constructor(config) { | ||
// data/models.ts | ||
// src/data/models.ts | ||
var OPEN_AI = "OpenAI"; | ||
@@ -698,3 +698,3 @@ var ANTHROPIC = "Anthropic"; | ||
// utils/get-llm-api-key.ts | ||
// src/utils/get-llm-api-key.ts | ||
function getLLMApiKey(modelProvider) { | ||
@@ -725,3 +725,3 @@ switch (true) { | ||
// utils/is-prod.ts | ||
// src/utils/is-prod.ts | ||
var FORCE_PROD = false; | ||
@@ -737,3 +737,3 @@ var TEST_PROD_LOCALLY = FORCE_PROD; | ||
// utils/to-old-pipe-format.ts | ||
// src/utils/to-old-pipe-format.ts | ||
function toOldPipeFormat(newFormat) { | ||
@@ -791,3 +791,3 @@ const [providerString, modelName] = newFormat.model.split(":"); | ||
// utils/local-server-running.ts | ||
// src/utils/local-server-running.ts | ||
async function isLocalServerRunning() { | ||
@@ -827,3 +827,3 @@ try { | ||
// pipes/pipes.ts | ||
// src/pipes/pipes.ts | ||
var Pipe = class { | ||
@@ -830,0 +830,0 @@ constructor(options) { |
@@ -20,3 +20,3 @@ "use strict"; | ||
// helpers/index.ts | ||
// src/helpers/index.ts | ||
var helpers_exports = {}; | ||
@@ -30,3 +30,3 @@ __export(helpers_exports, { | ||
// helpers/stream.ts | ||
// src/helpers/stream.ts | ||
var import_ChatCompletionStream = require("openai/lib/ChatCompletionStream"); | ||
@@ -33,0 +33,0 @@ var fromReadableStream = (readableStream) => { |
{ | ||
"name": "@baseai/core", | ||
"description": "The Web AI Framework's core - BaseAI.dev", | ||
"version": "0.9.15-snapshot.11", | ||
"version": "0.9.15-snapshot.12", | ||
"license": "Apache-2.0", | ||
@@ -12,5 +12,5 @@ "sideEffects": false, | ||
"dist/**", | ||
"react/dist/**", | ||
"pipes/dist/**", | ||
"helpers/dist/**", | ||
"react/dist/**", | ||
"CHANGELOG.md" | ||
@@ -17,0 +17,0 @@ ], |
@@ -20,3 +20,3 @@ "use strict"; | ||
// pipes/index.ts | ||
// src/pipes/index.ts | ||
var pipes_exports = {}; | ||
@@ -37,6 +37,6 @@ __export(pipes_exports, { | ||
// common/request.ts | ||
// src/common/request.ts | ||
var import_streaming = require("openai/streaming"); | ||
// common/errors.ts | ||
// src/common/errors.ts | ||
var APIError = class _APIError extends Error { | ||
@@ -155,3 +155,3 @@ constructor(status, error, message, headers) { | ||
// common/request.ts | ||
// src/common/request.ts | ||
var Request = class { | ||
@@ -291,3 +291,3 @@ constructor(config) { | ||
// data/models.ts | ||
// src/data/models.ts | ||
var OPEN_AI = "OpenAI"; | ||
@@ -681,3 +681,3 @@ var ANTHROPIC = "Anthropic"; | ||
// utils/get-llm-api-key.ts | ||
// src/utils/get-llm-api-key.ts | ||
function getLLMApiKey(modelProvider) { | ||
@@ -708,3 +708,3 @@ switch (true) { | ||
// utils/is-prod.ts | ||
// src/utils/is-prod.ts | ||
var FORCE_PROD = false; | ||
@@ -720,3 +720,3 @@ var TEST_PROD_LOCALLY = FORCE_PROD; | ||
// utils/to-old-pipe-format.ts | ||
// src/utils/to-old-pipe-format.ts | ||
function toOldPipeFormat(newFormat) { | ||
@@ -774,3 +774,3 @@ const [providerString, modelName] = newFormat.model.split(":"); | ||
// utils/local-server-running.ts | ||
// src/utils/local-server-running.ts | ||
async function isLocalServerRunning() { | ||
@@ -810,3 +810,3 @@ try { | ||
// pipes/pipes.ts | ||
// src/pipes/pipes.ts | ||
var Pipe = class { | ||
@@ -813,0 +813,0 @@ constructor(options) { |
@@ -21,3 +21,3 @@ 'use client' | ||
// react/index.ts | ||
// src/react/index.ts | ||
var react_exports = {}; | ||
@@ -29,7 +29,7 @@ __export(react_exports, { | ||
// react/use-pipe.ts | ||
// src/react/use-pipe.ts | ||
var import_react = require("react"); | ||
var import_zod = require("zod"); | ||
// helpers/stream.ts | ||
// src/helpers/stream.ts | ||
var import_ChatCompletionStream = require("openai/lib/ChatCompletionStream"); | ||
@@ -43,3 +43,3 @@ var fromReadableStream = (readableStream) => { | ||
// utils/is-prod.ts | ||
// src/utils/is-prod.ts | ||
var FORCE_PROD = false; | ||
@@ -52,3 +52,3 @@ var TEST_PROD_LOCALLY = FORCE_PROD; | ||
// react/use-pipe.ts | ||
// src/react/use-pipe.ts | ||
var uuidSchema = import_zod.z.string().uuid(); | ||
@@ -55,0 +55,0 @@ function usePipe({ |
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is not supported yet
Sorry, the diff of this file is not supported yet
418265