Azure Cosmos DB client library for JavaScript/TypeScript
Azure Cosmos DB is a globally distributed, multi-model database service that supports document, key-value, wide-column, and graph databases. This package is intended for JavaScript/TypeScript applications to interact with SQL API databases and the JSON documents they contain:
- Create Cosmos DB databases and modify their settings
- Create and modify containers to store collections of JSON documents
- Create, read, update, and delete the items (JSON documents) in your containers
- Query the documents in your database using SQL-like syntax
Key links:
Getting started
Prerequisites
Azure Subscription and Cosmos DB SQL API Account
You must have an Azure Subscription, and a Cosmos DB account (SQL API) to use this package.
If you need a Cosmos DB SQL API account, you can use the Azure Cloud Shell to create one with this Azure CLI command:
az cosmosdb create --resource-group <resource-group-name> --name <cosmos-database-account-name>
Or you can create an account in the Azure Portal
NodeJS
This package is distributed via npm which comes preinstalled with NodeJS using an LTS version.
CORS
You need to set up Cross-Origin Resource Sharing (CORS) rules for your Cosmos DB account if you need to develop for browsers. Follow the instructions in the linked document to create new CORS rules for your Cosmos DB.
Install this package
npm install @azure/cosmos
Get Account Credentials
You will need your Cosmos DB Account Endpoint and Key. You can find these in the Azure Portal or use the Azure CLI snippet below. The snippet is formatted for the Bash shell.
az cosmosdb show --resource-group <your-resource-group> --name <your-account-name> --query documentEndpoint --output tsv
az cosmosdb keys list --resource-group <your-resource-group> --name <your-account-name> --query primaryMasterKey --output tsv
Create an instance of CosmosClient
Interaction with Cosmos DB starts with an instance of the CosmosClient class
const { CosmosClient } = require("@azure/cosmos");
const endpoint = "https://your-account.documents.azure.com";
const key = "<database account masterkey>";
const client = new CosmosClient({ endpoint, key });
async function main() {
}
main().catch((error) => {
console.error(error);
});
For simplicity we have included the key
and endpoint
directly in the code but you will likely want to load these from a file not in source control using a project such as dotenv or loading from environment variables
In production environments, secrets like keys should be stored in Azure Key Vault
Key concepts
Once you've initialized a CosmosClient, you can interact with the primary resource types in Cosmos DB:
-
Database: A Cosmos DB account can contain multiple databases. When you create a database, you specify the API you'd like to use when interacting with its documents: SQL, MongoDB, Gremlin, Cassandra, or Azure Table. Use the Database object to manage its containers.
-
Container: A container is a collection of JSON documents. You create (insert), read, update, and delete items in a container by using methods on the Container object.
-
Item: An Item is a JSON document stored in a container. Each Item must include an id
key with a value that uniquely identifies the item within the container. If you do not provide an id
, the SDK will generate one automatically.
For more information about these resources, see Working with Azure Cosmos databases, containers and items.
Examples
The following sections provide several code snippets covering some of the most common Cosmos DB tasks, including:
Create a database
After authenticating your CosmosClient, you can work with any resource in the account. The code snippet below creates a NOSQL API database.
const { database } = await client.databases.createIfNotExists({ id: "Test Database" });
console.log(database.id);
Create a container
This example creates a container with default settings
const { container } = await database.containers.createIfNotExists({ id: "Test Database" });
console.log(container.id);
Using Partition Keys
This example shows various types of partition Keys supported.
await container.item("id", "1").read();
await container.item("id", 2).read();
await container.item("id", true).read();
await container.item("id", {}).read();
await container.item("id", undefined).read();
await container.item("id", null).read();
If the Partition Key consists of a single value, it could be supplied either as a literal value, or an array.
await container.item("id", "1").read();
await container.item("id", ["1"]).read();
If the Partition Key consists of more than one values, it should be supplied as an array.
await container.item("id", ["a", "b"]).read();
await container.item("id", ["a", 2]).read();
await container.item("id", [{}, {}]).read();
await container.item("id", ["a", {}]).read();
await container.item("id", [2, null]).read();
Insert items
To insert items into a container, pass an object containing your data to Items.upsert. The Azure Cosmos DB service requires each item has an id
key. If you do not provide one, the SDK will generate an id
automatically.
This example inserts several items into the container
const cities = [
{ id: "1", name: "Olympia", state: "WA", isCapitol: true },
{ id: "2", name: "Redmond", state: "WA", isCapitol: false },
{ id: "3", name: "Chicago", state: "IL", isCapitol: false }
];
for (const city of cities) {
await container.items.create(city);
}
Read an item
To read a single item from a container, use Item.read. This is a less expensive operation than using SQL to query by id
.
await container.item("1", "1").read();
CRUD on Container with hierarchical partition key
Create a Container with hierarchical partition key
const containerDefinition = {
id: "Test Database",
partitionKey: {
paths: ["/name", "/address/zip"],
version: PartitionKeyDefinitionVersion.V2,
kind: PartitionKeyKind.MultiHash,
},
}
const { container } = await database.containers.createIfNotExists(containerDefinition);
console.log(container.id);
Insert an item with hierarchical partition key defined as - ["/name", "/address/zip"]
const item = {
id: "1",
name: 'foo',
address: {
zip: 100
},
active: true
}
await container.items.create(item);
To read a single item from a container with hierarchical partition key defined as - ["/name", "/address/zip"],
await container.item("1", ["foo", 100]).read();
Query an item with hierarchical partition key with hierarchical partition key defined as - ["/name", "/address/zip"],
const { resources } = await container.items
.query("SELECT * from c WHERE c.active = true", {
partitionKey: ["foo", 100],
})
.fetchAll();
for (const item of resources) {
console.log(`${item.name}, ${item.address.zip} `);
}
Delete an item
To delete items from a container, use Item.delete.
await container.item("1").delete();
Query the database
A Cosmos DB SQL API database supports querying the items in a container with Items.query using SQL-like syntax:
const { resources } = await container.items
.query("SELECT * from c WHERE c.isCapitol = true")
.fetchAll();
for (const city of resources) {
console.log(`${city.name}, ${city.state} is a capitol `);
}
Perform parameterized queries by passing an object containing the parameters and their values to Items.query:
const { resources } = await container.items
.query({
query: "SELECT * from c WHERE c.isCapitol = @isCapitol",
parameters: [{ name: "@isCapitol", value: true }]
})
.fetchAll();
for (const city of resources) {
console.log(`${city.name}, ${city.state} is a capitol `);
}
For more information on querying Cosmos DB databases using the SQL API, see Query Azure Cosmos DB data with SQL queries.
Change Feed Pull Model
Change feed can be fetched for a partition key, a feed range or an entire container.
To process the change feed, create an instance of ChangeFeedPullModelIterator
. When you initially create ChangeFeedPullModelIterator
, you must specify a required changeFeedStartFrom
value inside the ChangeFeedIteratorOptions
which consists of both the starting position for reading changes and the resource(a partition key or a FeedRange) for which changes are to be fetched. You can optionally use maxItemCount
in ChangeFeedIteratorOptions
to set the maximum number of items received per page.
Note: If no changeFeedStartFrom
value is specified, then changefeed will be fetched for an entire container from Now().
There are four starting positions for change feed:
const options = {
changeFeedStartFrom: ChangeFeedStartFrom.Beginning(),
};
const iterator = container.getChangeFeedIterator(options);
const time = new Date("2023/09/11");
const options = {
changeFeedStartFrom: ChangeFeedStartFrom.Time(time),
};
const options = {
changeFeedStartFrom: ChangeFeedStartFrom.Now(),
};
const continuationToken = "some continuation token recieved from previous request";
const options = {
changeFeedStartFrom: ChangeFeedStartFrom.Continuation(continuationToken),
};
Here's an example of fetching change feed for a partition key
const partitionKey = "some-partition-Key-value";
const options = {
changeFeedStartFrom: ChangeFeedStartFrom.Beginning(partitionKey),
};
const iterator = container.items.getChangeFeedIterator(options);
while (iterator.hasMoreResults) {
const response = await iterator.readNext();
}
Because the change feed is effectively an infinite list of items that encompasses all future writes and updates, the value of hasMoreResults
is always true
. When you try to read the change feed and there are no new changes available, you receive a response with NotModified
status.
More detailed usage guidelines and examples of change feed can be found here.
Error Handling
The SDK generates various types of errors that can occur during an operation.
ErrorResponse
is thrown if the response of an operation returns an error code of >=400.TimeoutError
is thrown if Abort is called internally due to timeout.AbortError
is thrown if any user passed signal caused the abort.RestError
is thrown in case of failure of underlying system call due to network issues.- Errors generated by any devDependencies. For Eg.
@azure/identity
package could throw CredentialUnavailableError
.
Following is an example for handling errors of type ErrorResponse
, TimeoutError
, AbortError
, and RestError
.
try {
} catch (err) {
if (err instanceof ErrorResponse) {
} else if (err instanceof RestError) {
}
else {
}
}
It's important to properly handle these errors to ensure that your application can gracefully recover from any failures and continue functioning as expected. More details about some of these errors and their possible solutions can be found here.
Troubleshooting
General
When you interact with Cosmos DB errors returned by the service correspond to the same HTTP status codes returned for REST API requests:
HTTP Status Codes for Azure Cosmos DB
Conflicts
For example, if you try to create an item using an id
that's already in use in your Cosmos DB database, a 409
error is returned, indicating the conflict. In the following snippet, the error is handled gracefully by catching the exception and displaying additional information about the error.
try {
await containers.items.create({ id: "existing-item-id" });
} catch (error) {
if (error.code === 409) {
console.log("There was a conflict with an existing item");
}
}
Transpiling
The Azure SDKs are designed to support ES5 JavaScript syntax and LTS versions of Node.js. If you need support for earlier JavaScript runtimes such as Internet Explorer or Node 6, you will need to transpile the SDK code as part of your build process.
Handle transient errors with retries
While working with Cosmos DB, you might encounter transient failures caused by rate limits enforced by the service, or other transient problems like network outages. For information about handling these types of failures, see Retry pattern in the Cloud Design Patterns guide, and the related Circuit Breaker pattern.
Logging
Enabling logging may help uncover useful information about failures. In order to see a log of HTTP requests and responses, set the AZURE_LOG_LEVEL
environment variable to info
. Alternatively, logging can be enabled at runtime by calling setLogLevel
in the @azure/logger
. While using AZURE_LOG_LEVEL
make sure to set it before logging library is initialized.
Ideally pass it through command line, if using libraries like dotenv
make sure such libraries are initialized before logging library.
const { setLogLevel } = require("@azure/logger");
setLogLevel("info");
For more detailed instructions on how to enable logs, you can look at the @azure/logger package docs.
Diagnostics
Cosmos Diagnostics feature provides enhanced insights into all your client operations. A CosmosDiagnostics object is added to response of all client operations. such as
- Point look up operation reponse -
item.read()
, container.create()
, database.delete()
- Query operation reponse -
queryIterator.fetchAll()
, - Bulk and Batch operations -
item.batch()
. - Error/Exception response objects.
A CosmosDiagnostics object is added to response of all client operations.
There are 3 Cosmos Diagnostic levels, info, debug and debug-unsafe. Where only info is meant for production systems and debug and debug-unsafe are meant to be used during development and debugging, since they consume significantly higher resources. Cosmos Diagnostic level can be set in 2 ways
const client = new CosmosClient({ endpoint, key, diagnosticLevel: CosmosDbDiagnosticLevel.debug });
- Using environment variables. (Diagnostic level set by Environment variable has higher priority over setting it through client options.)
export AZURE_COSMOSDB_DIAGNOSTICS_LEVEL="debug"
Cosmos Diagnostic has three members
-
ClientSideRequestStatistics Type: Contains aggregates diagnostic details, including metadata lookups, retries, endpoints contacted, and request and response statistics like payload size and duration. (is always collected, can be used in production systems.)
-
DiagnosticNode: Is a tree-like structure that captures detailed diagnostic information. Similar to har
recording present in browsers. This feature is disabled by default and is intended for debugging non-production environments only. (collected at diagnostic level debug and debug-unsafe)
-
ClientConfig: Captures essential information related to client's configuration settings during client initialization. (collected at diagnostic level debug and debug-unsafe)
Please make sure to never set diagnostic level to debug-unsafe
in production environment, since it this level CosmosDiagnostics
captures request and response payloads and if you choose to log it (it is by default logged by @azure/logger at verbose
level). These payloads might get captured in your log sinks.
Consuming Diagnostics
- Since
diagnostics
is added to all Response objects. You could programatically access CosmosDiagnostic
as follows.
const { container, diagnostics: containerCreateDiagnostic } =
await database.containers.createIfNotExists({
id: containerId,
partitionKey: {
paths: ["/key1"],
},
});
const operations: OperationInput[] = [
{
operationType: BulkOperationType.Create,
resourceBody: { id: 'A', key: "A", school: "high" },
},
];
const response = await container.items.batch(operations, "A");
const queryIterator = container.items.query("select * from c");
const { resources, diagnostics } = await queryIterator.fetchAll();
try {
} catch (err) {
const diagnostics = err.diagnostics
}
- You could also log
diagnostics
using @azure/logger
, diagnostic is always logged using @azure/logger
at verbose
level. So if you set Diagnostic level to debug
or debug-unsafe
and @azure/logger
level to verbose
, diagnostics
will be logged.
Next steps
More sample code
Several samples are available to you in the SDK's GitHub repository. These samples provide example code for additional scenarios commonly encountered while working with Cosmos DB:
- Database Operations
- Container Operations
- Item Operations
- Configuring Indexing
- Reading a container Change Feed
- Stored Procedures
- Changing Database/Container throughput settings
- Multi Region Write Operations
Limitations
Currently the features below are not supported. For alternatives options, check the Workarounds section below.
Data Plane Limitations:
- Queries with COUNT from a DISTINCT subquery
- Direct TCP Mode access
- Aggregate cross-partition queries, like sorting, counting, and distinct, don't support continuation tokens. Streamable queries, like SELECT * FROM WHERE , support continuation tokens. See the "Workaround" section for executing non-streamable queries without a continuation token.
- Change Feed: Processor
- Change Feed: Read multiple partitions key values
- Change Feed pull model support for partial hierarchical partition keys #27059
- Cross-partition ORDER BY for mixed types
Control Plane Limitations:
- Get CollectionSizeUsage, DatabaseUsage, and DocumentUsage metrics
- Create Geospatial Index
- Update Autoscale throughput
Workarounds
Continuation token for cross partitions queries
You can achieve cross partition queries with continuation token support by using
Side car pattern.
This pattern can also enable applications to be composed of heterogeneous components and technologies.
Executing non-stremable cross-partition query
To execute non-streamable queries without the use of continuation tokens, you can create a query iterator with the required query specification and options. The following sample code demonstrates how to use a query iterator to fetch all results without the need for a continuation token:
const querySpec = {
query: "SELECT c.status, COUNT(c.id) AS count FROM c GROUP BY c.status",
};
const queryOptions = {
maxItemCount: 10,
enableCrossPartitionQuery: true,
};
const querIterator = await container.items.query(querySpec, queryOptions);
while (querIterator.hasMoreResults()) {
const { resources: result } = await querIterator.fetchNext();
}
This approach can also be used for streamable queries.
Control Plane operations
Typically, you can use Azure Portal, Azure Cosmos DB Resource Provider REST API, Azure CLI or PowerShell for the control plane unsupported limitations.
Additional documentation
For more extensive documentation on the Cosmos DB service, see the Azure Cosmos DB documentation on docs.microsoft.com.
Useful links
Contributing
If you'd like to contribute to this library, please read the contributing guide to learn more about how to build and test the code.