Update (2023-02-02)
Trying to use text-chat-davinci-002-20230126
with the OpenAI API now returns a 404 error. Someone has already found the new model name, but they are unwilling to share at this time. I will update this repository once I find the new model. If you have any leads, please open an issue or a pull request.
In the meantime, I've added support for models like text-davinci-003
, which you can use as a drop-in replacement. Keep in mind that text-davinci-003
is not as good as text-chat-davinci-002
(which is trained via RHLF and fine-tuned to be a conversational AI), though results are still very good. Please note that using text-davinci-003
will cost you credits ($).
Discord user @pig#8932 has found a working text-chat-davinci-002
model, text-chat-davinci-002-20221122
. I've updated the library to use this model.
ChatGPT API
A ChatGPT implementation using the official ChatGPT model via OpenAI's API.
This is an implementation of ChatGPT using the official ChatGPT raw model, text-chat-davinci-002
. This model name text-chat-davinci-002-20230126
was briefly leaked while I was inspecting the network requests made by the official ChatGPT website, and I discovered that it works with the OpenAI API. Usage of this model currently does not cost any credits.
As far as I'm aware, I was the first one who discovered this, and usage of the model has since been implemented in libraries like acheong08/ChatGPT.
The previous version of this library that used transitive-bullshit/chatgpt-api is still available on the archive/old-version
branch.
By itself, the model does not have any conversational support, so this library uses a cache to store conversations and pass them to the model as context. This allows you to have persistent conversations with ChatGPT in a nearly identical way to the official website.
Features
- Uses the official ChatGPT raw model,
text-chat-davinci-002-20221122
. - Includes an API server you can run to use ChatGPT in non-Node.js applications.
- Includes a
ChatGPTClient
class that you can use in your own Node.js applications. - Includes a CLI interface where you can chat with ChatGPT.
- Replicates chat threads from the official ChatGPT website (with conversation IDs and message IDs), with persistent conversations using Keyv.
- Conversations are stored in memory by default, but you can optionally install a storage adapter to persist conversations to a database.
- The
keyv-file
adapter is also included in this package, and can be used to store conversations in a JSON file if you're using the API server or CLI (see settings.example.js
).
- Supports configurable prompt prefixes, and custom names for the user and ChatGPT.
- In essence, this allows you to turn ChatGPT into a different character.
- This is currently only configurable on a global level, but I plan to add support for per-conversation customization.
Getting Started
Prerequisites
Usage
Module
npm i @waylaidwanderer/chatgpt-api
import ChatGPTClient from '@waylaidwanderer/chatgpt-api';
const clientOptions = {
modelOptions: {
model: 'text-chat-davinci-002-20221122',
},
debug: false,
};
const cacheOptions = {
};
const chatGptClient = new ChatGPTClient('OPENAI_API_KEY', clientOptions, cacheOptions);
const response = await chatGptClient.sendMessage('Hello!');
console.log(response);
const response2 = await chatGptClient.sendMessage('Write a poem about cats.', { conversationId: response.conversationId, parentMessageId: response.messageId });
console.log(response2.response);
const response3 = await chatGptClient.sendMessage('Now write it in French.', { conversationId: response2.conversationId, parentMessageId: response2.messageId });
console.log(response3.response);
API Server
You can install the package using
npm i -g @waylaidwanderer/chatgpt-api
then run it using
chatgpt-api
.
This takes an optional --settings=<path_to_settings.js>
parameter, or looks for settings.js
in the current directory if not set, with the following contents:
module.exports = {
openaiApiKey: process.env.OPENAI_API_KEY || '',
chatGptClient: {
modelOptions: {
model: 'text-chat-davinci-002-20221122',
},
debug: false,
},
cacheOptions: {},
apiOptions: {
port: process.env.API_PORT || 3000,
host: process.env.API_HOST || 'localhost',
},
storageFilePath: process.env.STORAGE_FILE_PATH || './cache.json',
};
Alternatively, you can install and run the package locally:
- Clone this repository
- Install dependencies with
npm install
- Rename
settings.example.js
to settings.js
in the root directory and change the settings where required. - Start the server using
npm start
or npm run server
To start a conversation with ChatGPT, send a POST request to the server's /conversation
endpoint with a JSON body in the following format:
{
"message": "Hello, how are you today?",
"conversationId": "your-conversation-id (optional)",
"parentMessageId": "your-parent-message-id (optional)"
}
The server will return a JSON object containing ChatGPT's response:
{
"response": "I'm doing well, thank you! How are you?",
"conversationId": "your-conversation-id",
"messageId": "response-message-id"
}
If the request is unsuccessful, the server will return a JSON object with an error message and a status code of 503.
If there was an error sending the message to ChatGPT:
{
"error": "There was an error communicating with ChatGPT."
}
CLI
Install the package using the same instructions as the API server.
If installed globally:
chatgpt-cli
If installed locally:
npm run cli
ChatGPT's responses are automatically copied to your clipboard, so you can paste them into other applications.
Caveats
Since text-chat-davinci-002-20221122
is ChatGPT's raw model, I had to do my best to replicate the way the official ChatGPT website uses it. After extensive testing and comparing responses, I believe that the model used by ChatGPT has some additional fine-tuning.
This means my implementation or the raw model may not behave exactly the same in some ways:
-
Conversations are not tied to any user IDs, so if that's important to you, you should implement your own user ID system.
-
ChatGPT's model parameters (temperature, frequency penalty, etc.) are unknown, so I set some defaults that I thought would be reasonable.
-
Conversations are limited to roughly the last 3000 tokens, so earlier messages may be forgotten during longer conversations.
- This works in a similar way to ChatGPT, except I'm pretty sure they have some additional way of retrieving context from earlier messages when needed (which can probably be achieved with embeddings, but I consider that out-of-scope for now).
-
It is well known that, as part of the fine-tuning, ChatGPT had the following preamble:
"You are ChatGPT, a large language model trained by OpenAI. You answer as concisely as possible for each response (e.g. don’t be verbose). It is very important that you answer as concisely as possible, so please remember this. If you are generating a list, do not have too many items. Keep the number of items short.
Knowledge cutoff: 2021-09
Current date: 2023-01-31"
As OpenAI updates ChatGPT, this preamble may also change. The default prompt prefix in my implementation attempts to replicate a similar behavior to the current ChatGPT model.
Contributing
If you'd like to contribute to this project, please create a pull request with a detailed description of your changes.
License
This project is licensed under the MIT License.