OpenAI Node API Library
The OpenAI Node library provides convenient access to the OpenAI REST API from applications written in server-side JavaScript.
It includes TypeScript definitions for all request params and response fields.
⚠️ Important note: this library is meant for server-side usage only, as using it in client-side browser code will expose your secret API key. See here for more details.
Documentation
To learn how to use the OpenAI API, check out our API Reference and Documentation.
Installation
npm install --save openai
yarn add openai
Usage
import OpenAI from 'openai';
const openai = new OpenAI({
apiKey: 'my api key',
});
async function main() {
const completion = await openai.chat.completions.create({
messages: [{ role: 'user', content: 'Say this is a test' }],
model: 'gpt-3.5-turbo',
});
console.log(completion.choices);
}
main();
Streaming Responses
We provide support for streaming responses using Server Side Events (SSE).
import OpenAI from 'openai';
const openai = new OpenAI();
async function main() {
const completion = await openai.chat.completions.create({
model: 'gpt-4',
messages: [{ role: 'user', content: 'Say this is a test' }],
stream: true,
});
for await (const part of stream) {
process.stdout.write(part.choices[0]?.text || '');
}
}
main();
If you need to cancel a stream, you can break
from the loop
or call stream.controller.abort()
.
Usage with TypeScript
Importing, instantiating, and interacting with the library are the same as above.
If you like, you may reference our types directly:
import OpenAI from 'openai';
const openai = new OpenAI({
apiKey: 'my api key',
});
async function main() {
const params: OpenAI.Chat.CompletionCreateParams = {
messages: [{ role: 'user', content: 'Say this is a test' }],
model: 'gpt-3.5-turbo',
};
const completion: OpenAI.Chat.ChatCompletion = await openai.chat.completions.create(params);
}
main();
Documentation for each method, request param, and response field are available in docstrings and will appear on hover in most modern editors.
File Uploads
Request parameters that correspond to file uploads can be passed in many different forms:
File
(or an object with the same structure)- a
fetch
Response
(or an object with the same structure) - an
fs.ReadStream
- the return value of our
toFile
helper
import fs from 'fs';
import fetch from 'node-fetch';
import OpenAI, { toFile } from 'openai';
const openai = new OpenAI();
await openai.files.create({ file: fs.createReadStream('input.jsonl'), purpose: 'fine-tune' });
await openai.files.create({ file: new File(['my bytes'], 'input.jsonl'), purpose: 'fine-tune' });
await openai.files.create({ file: await fetch('https://somesite/input.jsonl'), purpose: 'fine-tune' });
await openai.files.create({
file: await toFile(Buffer.from('my bytes'), 'input.jsonl'),
purpose: 'fine-tune',
});
await openai.files.create({
file: await toFile(new Uint8Array([0, 1, 2]), 'input.jsonl'),
purpose: 'fine-tune',
});
Handling errors
When the library is unable to connect to the API,
or if the API returns a non-success status code (i.e., 4xx or 5xx response),
a subclass of APIError
will be thrown:
async function main() {
const fineTune = await openai.fineTunes
.create({ training_file: 'file-XGinujblHPwGLSztz8cPS8XY' })
.catch((err) => {
if (err instanceof OpenAI.APIError) {
console.log(err.status);
console.log(err.name);
console.log(err.headers);
} else {
throw err;
}
});
}
main();
Error codes are as followed:
Status Code | Error Type |
---|
400 | BadRequestError |
401 | AuthenticationError |
403 | PermissionDeniedError |
404 | NotFoundError |
422 | UnprocessableEntityError |
429 | RateLimitError |
>=500 | InternalServerError |
N/A | APIConnectionError |
Retries
Certain errors will be automatically retried 2 times by default, with a short exponential backoff.
Connection errors (for example, due to a network connectivity problem), 409 Conflict, 429 Rate Limit,
and >=500 Internal errors will all be retried by default.
You can use the maxRetries
option to configure or disable this:
const openai = new OpenAI({
maxRetries: 0,
});
await openai.chat.completions.create({ messages: [{ role: 'user', content: 'How can I get the name of the current day in Node.js?' }], model: 'gpt-3.5-turbo' }, {
maxRetries: 5,
});
Timeouts
Requests time out after 60 seconds by default. You can configure this with a timeout
option:
const openai = new OpenAI({
timeout: 20 * 1000,
});
await openai.chat.completions.create({ messages: [{ role: 'user', content: 'How can I list all files in a directory using Python?' }], model: 'gpt-3.5-turbo' }, {
timeout: 5 * 1000,
});
On timeout, an APIConnectionTimeoutError
is thrown.
Note that requests which time out will be retried twice by default.
Configuring an HTTP(S) Agent (e.g., for proxies)
By default, this library uses a stable agent for all http/https requests to reuse TCP connections, eliminating many TCP & TLS handshakes and shaving around 100ms off most requests.
If you would like to disable or customize this behavior, for example to use the API behind a proxy, you can pass an httpAgent
which is used for all requests (be they http or https), for example:
import http from 'http';
import HttpsProxyAgent from 'https-proxy-agent';
const openai = new OpenAI({
httpAgent: new HttpsProxyAgent(process.env.PROXY_URL),
});
await openai.models.list({
baseURL: 'http://localhost:8080/test-api',
httpAgent: new http.Agent({ keepAlive: false }),
})
Status
This package is in beta. Its internals and interfaces are not stable
and subject to change without a major semver bump;
please reach out if you rely on any undocumented behavior.
We are keen for your feedback; please open an issue with questions, bugs, or suggestions.
Requirements
The following runtimes are supported:
- Node.js 16 LTS or later (non-EOL) versions.
- Deno v1.28.0 or higher (experimental).
Use
import OpenAI from "npm:openai"
.
If you are interested in other runtime environments, please open or upvote an issue on GitHub.