Huge News!Announcing our $40M Series B led by Abstract Ventures.Learn More
Socket
Sign inDemoInstall
Socket

openai-streams

Package Overview
Dependencies
Maintainers
1
Versions
105
Alerts
File Explorer

Advanced tools

Socket logo

Install Socket

Detect and block malicious and high-risk dependencies

Install

openai-streams

Tools for working with OpenAI streams in Node.js and TypeScript.

  • 1.0.22-canary.2
  • npm
  • Socket score

Version published
Maintainers
1
Created
Source

OpenAI Streams

Github | NPM | Docs

This library returns OpenAI API responses as streams only. Non-stream endpoints like edits etc. are simply a stream with only one chunk update.

It simplifies the following:

  1. Prioritizing streaming and type inference.
  2. Auto-loads OPENAI_API_KEY from process.env.
  3. Uses the same function for all endpoints, and switches the type based on the OpenAI(endpoint, ...) signature.

Overall, the library aims to make it as simple to call the API as possible and stream updates in.

Installation

yarn add openai-streams

# or

npm i --save openai-streams

Usage

  1. Set the OPENAI_API_KEY env variable (or pass the { apiKey } option).

    The library will throw if it cannot find an API key. Your program will load this at runtime from process.env.OPENAI_API_KEY by default, but you may override this with the { apiKey } option.

    IMPORTANT: For security, you should only load this from a process.env variable.

    await OpenAI(
      "completions", 
      {/* params */}, 
      { apiKey: process.env.MY_SECRET_API_KEY }
    )
    
  2. Call the API via await OpenAI(endpoint, params).

    The params type will be inferred based on the endpoint you provide, i.e. for the "edits" endpoint, import('openai').CreateEditRequest will be enforced.

Example: Consuming streams in Next.js Edge functions
import { OpenAI } from "openai-streams";

export default async function handler() {
  const stream = await OpenAI(
    "completions",
    {
      model: "text-davinci-003",
      prompt: "Write a sentence.\n\n",
      max_tokens: 100
    },
  );

  return new Response(stream);
}

export const config = {
  runtime: "edge"
};

See the example in example/src/pages/api/hello.ts. See also src/pages/api/demo.ts in nextjs-openai.

Notes

  1. Internally, streams are often manipulated using generators via for await (const chunk of yieldStream(stream)) { ... }. We recommend following this pattern if you find it intuitive.

Keywords

FAQs

Package last updated on 23 Feb 2023

Did you know?

Socket

Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.

Install

Related posts

SocketSocket SOC 2 Logo

Product

  • Package Alerts
  • Integrations
  • Docs
  • Pricing
  • FAQ
  • Roadmap
  • Changelog

Packages

npm

Stay in touch

Get open source security insights delivered straight into your inbox.


  • Terms
  • Privacy
  • Security

Made with ⚡️ by Socket Inc