Huge News!Announcing our $40M Series B led by Abstract Ventures.Learn More β†’
Socket
Sign inDemoInstall
Socket

carbon_ruby_sdk

Package Overview
Dependencies
Maintainers
1
Alerts
File Explorer

Advanced tools

Socket logo

Install Socket

Detect and block malicious and high-risk dependencies

Install

carbon_ruby_sdk

  • 0.2.53
  • Rubygems
  • Socket score

Version published
Maintainers
1
Created
Source

Visit Carbon

Carbon

Connect external data to LLMs, no matter the source.

npm

Table of Contents

Installation

Add to Gemfile:

gem 'carbon_ruby_sdk', '~> 0.2.53'

Getting Started

require 'carbon_ruby_sdk'

# 1) Get an access token for a customer
configuration = Carbon::Configuration.new
configuration.api_key = "YOUR_API_KEY"
configuration.customer_id = "YOUR_CUSTOMER_ID"
carbon = Carbon::Client.new(configuration)

token = carbon.auth.get_access_token

# 2) Use the access token to authenticate moving forward
configuration = Carbon::Configuration.new
configuration.access_token = token.access_token
carbon = Carbon::Client.new(configuration)

# use SDK as usual
white_labeling = carbon.auth.get_white_labeling

Raw HTTP Response

To access the raw HTTP response, suffix any method with _with_http_info.

result = carbon.auth.get_access_token_with_http_info
p result.data # [TokenResponse] Deserialized data
p.result.status_code # [Integer] HTTP status code
p.result.headers # [Hash] HTTP headers
p.result.response # [Faraday::Response] Raw HTTP response

Reference

carbon.auth.get_access_token

Get Access Token

πŸ› οΈ Usage
result = carbon.auth.get_access_token
p result
πŸ”„ Return

TokenResponse

🌐 Endpoint

/auth/v1/access_token GET

πŸ”™ Back to Table of Contents


carbon.auth.get_white_labeling

Returns whether or not the organization is white labeled and which integrations are white labeled

:param current_user: the current user :param db: the database session :return: a WhiteLabelingResponse

πŸ› οΈ Usage
result = carbon.auth.get_white_labeling
p result
πŸ”„ Return

WhiteLabelingResponse

🌐 Endpoint

/auth/v1/white_labeling GET

πŸ”™ Back to Table of Contents


carbon.crm.get_account

Get Account

πŸ› οΈ Usage
result = carbon.crm.get_account(
  id: "id_example",
  data_source_id: 1,
  include_remote_data: false,
  includes: [
        "string_example"
    ],
)
p result
βš™οΈ Parameters
id: String
data_source_id: Integer
include_remote_data: Boolean
includes: Array<BaseIncludes>
πŸ”„ Return

Account

🌐 Endpoint

/integrations/data/crm/accounts/{id} GET

πŸ”™ Back to Table of Contents


carbon.crm.get_accounts

Get Accounts

πŸ› οΈ Usage
result = carbon.crm.get_accounts(
  data_source_id: 1,
  include_remote_data: false,
  next_cursor: "string_example",
  page_size: 1,
  order_dir: "asc",
  includes: [],
  filters: {
    },
  order_by: "created_at",
)
p result
βš™οΈ Parameters
data_source_id: Integer
include_remote_data: Boolean
next_cursor: String
page_size: Integer
order_dir: OrderDirV2Nullable
includes: Array<BaseIncludes>
filters: AccountFilters
order_by: AccountsOrderByNullable
πŸ”„ Return

AccountResponse

🌐 Endpoint

/integrations/data/crm/accounts POST

πŸ”™ Back to Table of Contents


carbon.crm.get_contact

Get Contact

πŸ› οΈ Usage
result = carbon.crm.get_contact(
  id: "id_example",
  data_source_id: 1,
  include_remote_data: false,
  includes: [
        "string_example"
    ],
)
p result
βš™οΈ Parameters
id: String
data_source_id: Integer
include_remote_data: Boolean
includes: Array<BaseIncludes>
πŸ”„ Return

Contact

🌐 Endpoint

/integrations/data/crm/contacts/{id} GET

πŸ”™ Back to Table of Contents


carbon.crm.get_contacts

Get Contacts

πŸ› οΈ Usage
result = carbon.crm.get_contacts(
  data_source_id: 1,
  include_remote_data: false,
  next_cursor: "string_example",
  page_size: 1,
  order_dir: "asc",
  includes: [],
  filters: {
    },
  order_by: "created_at",
)
p result
βš™οΈ Parameters
data_source_id: Integer
include_remote_data: Boolean
next_cursor: String
page_size: Integer
order_dir: OrderDirV2Nullable
includes: Array<BaseIncludes>
filters: ContactFilters
order_by: ContactsOrderByNullable
πŸ”„ Return

ContactsResponse

🌐 Endpoint

/integrations/data/crm/contacts POST

πŸ”™ Back to Table of Contents


carbon.crm.get_lead

Get Lead

πŸ› οΈ Usage
result = carbon.crm.get_lead(
  id: "id_example",
  data_source_id: 1,
  include_remote_data: false,
  includes: [
        "string_example"
    ],
)
p result
βš™οΈ Parameters
id: String
data_source_id: Integer
include_remote_data: Boolean
includes: Array<BaseIncludes>
πŸ”„ Return

Lead

🌐 Endpoint

/integrations/data/crm/leads/{id} GET

πŸ”™ Back to Table of Contents


carbon.crm.get_leads

Get Leads

πŸ› οΈ Usage
result = carbon.crm.get_leads(
  data_source_id: 1,
  include_remote_data: false,
  next_cursor: "string_example",
  page_size: 1,
  order_dir: "asc",
  includes: [],
  filters: {
    },
  order_by: "created_at",
)
p result
βš™οΈ Parameters
data_source_id: Integer
include_remote_data: Boolean
next_cursor: String
page_size: Integer
order_dir: OrderDirV2Nullable
includes: Array<BaseIncludes>
filters: LeadFilters
order_by: LeadsOrderByNullable
πŸ”„ Return

LeadsResponse

🌐 Endpoint

/integrations/data/crm/leads POST

πŸ”™ Back to Table of Contents


carbon.crm.get_opportunities

Get Opportunities

πŸ› οΈ Usage
result = carbon.crm.get_opportunities(
  data_source_id: 1,
  include_remote_data: false,
  next_cursor: "string_example",
  page_size: 1,
  order_dir: "asc",
  includes: [],
  filters: {
        "status" => "WON",
    },
  order_by: "created_at",
)
p result
βš™οΈ Parameters
data_source_id: Integer
include_remote_data: Boolean
next_cursor: String
page_size: Integer
order_dir: OrderDirV2Nullable
includes: Array<BaseIncludes>
filters: OpportunityFilters
order_by: OpportunitiesOrderByNullable
πŸ”„ Return

OpportunitiesResponse

🌐 Endpoint

/integrations/data/crm/opportunities POST

πŸ”™ Back to Table of Contents


carbon.crm.get_opportunity

Get Opportunity

πŸ› οΈ Usage
result = carbon.crm.get_opportunity(
  id: "id_example",
  data_source_id: 1,
  include_remote_data: false,
  includes: [
        "string_example"
    ],
)
p result
βš™οΈ Parameters
id: String
data_source_id: Integer
include_remote_data: Boolean
includes: Array<BaseIncludes>
πŸ”„ Return

Opportunity

🌐 Endpoint

/integrations/data/crm/opportunities/{id} GET

πŸ”™ Back to Table of Contents


carbon.data_sources.add_tags

Add Data Source Tags

πŸ› οΈ Usage
result = carbon.data_sources.add_tags(
  tags: {},
  data_source_id: 1,
)
p result
βš™οΈ Parameters
tags: Object
data_source_id: Integer
πŸ”„ Return

OrganizationUserDataSourceAPI

🌐 Endpoint

/data_sources/tags/add POST

πŸ”™ Back to Table of Contents


carbon.data_sources.query

Data Sources

πŸ› οΈ Usage
result = carbon.data_sources.query(
  pagination: {
        "limit" => 10,
        "offset" => 0,
        "starting_id" => 0,
    },
  order_by: "created_at",
  order_dir: "desc",
  filters: {
        "source" => "GOOGLE_CLOUD_STORAGE",
    },
)
p result
βš™οΈ Parameters
pagination: Pagination
order_by: OrganizationUserDataSourceOrderByColumns
order_dir: OrderDir
filters: OrganizationUserDataSourceFilters
πŸ”„ Return

OrganizationUserDataSourceResponse

🌐 Endpoint

/data_sources POST

πŸ”™ Back to Table of Contents


carbon.data_sources.query_user_data_sources

User Data Sources

πŸ› οΈ Usage
result = carbon.data_sources.query_user_data_sources(
  pagination: {
        "limit" => 10,
        "offset" => 0,
        "starting_id" => 0,
    },
  order_by: "created_at",
  order_dir: "desc",
  filters: {
        "source" => "GOOGLE_CLOUD_STORAGE",
    },
)
p result
βš™οΈ Parameters
pagination: Pagination
order_by: OrganizationUserDataSourceOrderByColumns
order_dir: OrderDir
filters: OrganizationUserDataSourceFilters
πŸ”„ Return

OrganizationUserDataSourceResponse

🌐 Endpoint

/user_data_sources POST

πŸ”™ Back to Table of Contents


carbon.data_sources.remove_tags

Remove Data Source Tags

πŸ› οΈ Usage
result = carbon.data_sources.remove_tags(
  data_source_id: 1,
  tags_to_remove: [],
  remove_all_tags: false,
)
p result
βš™οΈ Parameters
data_source_id: Integer
tags_to_remove: Array<String>
remove_all_tags: Boolean
πŸ”„ Return

OrganizationUserDataSourceAPI

🌐 Endpoint

/data_sources/tags/remove POST

πŸ”™ Back to Table of Contents


carbon.data_sources.revoke_access_token

Revoke Access Token

πŸ› οΈ Usage
result = carbon.data_sources.revoke_access_token(
  data_source_id: 1,
)
p result
βš™οΈ Parameters
data_source_id: Integer
πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/revoke_access_token POST

πŸ”™ Back to Table of Contents


carbon.embeddings.get_documents

For pre-filtering documents, using tags_v2 is preferred to using tags (which is now deprecated). If both tags_v2 and tags are specified, tags is ignored. tags_v2 enables building complex filters through the use of "AND", "OR", and negation logic. Take the below input as an example:

{
    "OR": [
        {
            "key": "subject",
            "value": "holy-bible",
            "negate": false
        },
        {
            "key": "person-of-interest",
            "value": "jesus christ",
            "negate": false
        },
        {
            "key": "genre",
            "value": "religion",
            "negate": true
        }
        {
            "AND": [
                {
                    "key": "subject",
                    "value": "tao-te-ching",
                    "negate": false
                },
                {
                    "key": "author",
                    "value": "lao-tzu",
                    "negate": false
                }
            ]
        }
    ]
}

In this case, files will be filtered such that:

  1. "subject" = "holy-bible" OR
  2. "person-of-interest" = "jesus christ" OR
  3. "genre" != "religion" OR
  4. "subject" = "tao-te-ching" AND "author" = "lao-tzu"

Note that the top level of the query must be either an "OR" or "AND" array. Currently, nesting is limited to 3. For tag blocks (those with "key", "value", and "negate" keys), the following typing rules apply:

  1. "key" isn't optional and must be a string
  2. "value" isn't optional and can be any or list[any]
  3. "negate" is optional and must be true or false. If present and true, then the filter block is negated in the resulting query. It is false by default.

When querying embeddings, you can optionally specify the media_type parameter in your request. By default (if not set), it is equal to "TEXT". This means that the query will be performed over files that have been parsed as text (for now, this covers all files except image files). If it is equal to "IMAGE", the query will be performed over image files (for now, .jpg and .png files). You can think of this field as an additional filter on top of any filters set in file_ids and

When hybrid_search is set to true, a combination of keyword search and semantic search are used to rank and select candidate embeddings during information retrieval. By default, these search methods are weighted equally during the ranking process. To adjust the weight (or "importance") of each search method, you can use the hybrid_search_tuning_parameters property. The description for the different tuning parameters are:

  • weight_a: weight to assign to semantic search
  • weight_b: weight to assign to keyword search

You must ensure that sum(weight_a, weight_b,..., weight_n) for all n weights is equal to 1. The equality has an error tolerance of 0.001 to account for possible floating point issues.

In order to use hybrid search for a customer across a set of documents, two flags need to be enabled:

  1. Use the /modify_user_configuration endpoint to to enable sparse_vectors for the customer. The payload body for this request is below:
{
  "configuration_key_name": "sparse_vectors",
  "value": {
    "enabled": true
  }
}
  1. Make sure hybrid search is enabled for the documents across which you want to perform the search. For the /uploadfile endpoint, this can be done by setting the following query parameter: generate_sparse_vectors=true

Carbon supports multiple models for use in generating embeddings for files. For images, we support Vertex AI's multimodal model; for text, we support OpenAI's text-embedding-ada-002 and Cohere's embed-multilingual-v3.0. The model can be specified via the embedding_model parameter (in the POST body for /embeddings, and a query parameter in /uploadfile). If no model is supplied, the text-embedding-ada-002 is used by default. When performing embedding queries, embeddings from files that used the specified model will be considered in the query. For example, if files A and B have embeddings generated with OPENAI, and files C and D have embeddings generated with COHERE_MULTILINGUAL_V3, then by default, queries will only consider files A and B. If COHERE_MULTILINGUAL_V3 is specified as the embedding_model in /embeddings, then only files C and D will be considered. Make sure that the set of all files you want considered for a query have embeddings generated via the same model. For now, do not set VERTEX_MULTIMODAL as an embedding_model. This model is used automatically by Carbon when it detects an image file.

πŸ› οΈ Usage
result = carbon.embeddings.get_documents(
  query: "a",
  k: 1,
  tags: {
        "key": "string_example",
    },
  query_vector: [
        3.14
    ],
  file_ids: [
        1
    ],
  parent_file_ids: [
        1
    ],
  include_all_children: false,
  tags_v2: {
    },
  include_tags: true,
  include_vectors: true,
  include_raw_file: true,
  hybrid_search: true,
  hybrid_search_tuning_parameters: {
        "weight_a" => 0.5,
        "weight_b" => 0.5,
    },
  media_type: "TEXT",
  embedding_model: "OPENAI",
  include_file_level_metadata: false,
  high_accuracy: false,
  rerank: {
        "model" => "model_example",
    },
  file_types_at_source: [
        "string_example"
    ],
  exclude_cold_storage_files: false,
)
p result
βš™οΈ Parameters
query: String

Query for which to get related chunks and embeddings.

k: Integer

Number of related chunks to return.

tags: Hash<string, Tags1>

A set of tags to limit the search to. Deprecated and may be removed in the future.

query_vector: Array<Float>

Optional query vector for which to get related chunks and embeddings. It must have been generated by the same model used to generate the embeddings across which the search is being conducted. Cannot provide both query and query_vector.

file_ids: Array<Integer>

Optional list of file IDs to limit the search to

parent_file_ids: Array<Integer>

Optional list of parent file IDs to limit the search to. A parent file describes a file to which another file belongs (e.g. a folder)

include_all_children: Boolean

Flag to control whether or not to include all children of filtered files in the embedding search.

tags_v2: Object

A set of tags to limit the search to. Use this instead of tags, which is deprecated.

include_tags: Boolean

Flag to control whether or not to include tags for each chunk in the response.

include_vectors: Boolean

Flag to control whether or not to include embedding vectors in the response.

include_raw_file: Boolean

Flag to control whether or not to include a signed URL to the raw file containing each chunk in the response.

hybrid_search: Boolean

Flag to control whether or not to perform hybrid search.

hybrid_search_tuning_parameters: HybridSearchTuningParamsNullable
media_type: FileContentTypesNullable
embedding_model: EmbeddingGeneratorsNullable
include_file_level_metadata: Boolean

Flag to control whether or not to include file-level metadata in the response. This metadata will be included in the content_metadata field of each document along with chunk/embedding level metadata.

high_accuracy: Boolean

Flag to control whether or not to perform a high accuracy embedding search. By default, this is set to false. If true, the search may return more accurate results, but may take longer to complete.

rerank: RerankParamsNullable
file_types_at_source: Array<AutoSyncedSourceTypesPropertyInner>

Filter files based on their type at the source (for example help center tickets and articles)

exclude_cold_storage_files: Boolean

Flag to control whether or not to exclude files that are not in hot storage. If set to False, then an error will be returned if any filtered files are in cold storage.

πŸ”„ Return

DocumentResponseList

🌐 Endpoint

/embeddings POST

πŸ”™ Back to Table of Contents


carbon.embeddings.get_embeddings_and_chunks

Retrieve Embeddings And Content

πŸ› οΈ Usage
result = carbon.embeddings.get_embeddings_and_chunks(
  filters: {
        "user_file_id" => 1,
        "embedding_model" => "OPENAI",
    },
  pagination: {
        "limit" => 10,
        "offset" => 0,
        "starting_id" => 0,
    },
  order_by: "created_at",
  order_dir: "desc",
  include_vectors: false,
)
p result
βš™οΈ Parameters
filters: EmbeddingsAndChunksFilters
pagination: Pagination
order_by: EmbeddingsAndChunksOrderByColumns
order_dir: OrderDir
include_vectors: Boolean
πŸ”„ Return

EmbeddingsAndChunksResponse

🌐 Endpoint

/text_chunks POST

πŸ”™ Back to Table of Contents


carbon.embeddings.list

Retrieve Embeddings And Content V2

πŸ› οΈ Usage
result = carbon.embeddings.list(
  filters: {
        "include_all_children" => false,
        "non_synced_only" => false,
    },
  pagination: {
        "limit" => 10,
        "offset" => 0,
        "starting_id" => 0,
    },
  order_by: "created_at",
  order_dir: "desc",
  include_vectors: false,
)
p result
βš™οΈ Parameters
filters: OrganizationUserFilesToSyncFilters
pagination: Pagination
order_by: OrganizationUserFilesToSyncOrderByTypes
order_dir: OrderDir
include_vectors: Boolean
πŸ”„ Return

EmbeddingsAndChunksResponse

🌐 Endpoint

/list_chunks_and_embeddings POST

πŸ”™ Back to Table of Contents


carbon.embeddings.upload_chunks_and_embeddings

Upload Chunks And Embeddings

πŸ› οΈ Usage
result = carbon.embeddings.upload_chunks_and_embeddings(
  embedding_model: "OPENAI",
  chunks_and_embeddings: [
        {
            "file_id" => 1,
            "chunks_and_embeddings" => [
                {
                    "chunk_number" => 1,
                    "chunk" => "chunk_example",
                }
            ],
        }
    ],
  overwrite_existing: false,
  chunks_only: false,
  custom_credentials: {
        "key": {},
    },
)
p result
βš™οΈ Parameters
embedding_model: EmbeddingGenerators
chunks_and_embeddings: Array<SingleChunksAndEmbeddingsUploadInput>
overwrite_existing: Boolean
chunks_only: Boolean
custom_credentials: Hash<String, Object>
πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/upload_chunks_and_embeddings POST

πŸ”™ Back to Table of Contents


carbon.files.create_user_file_tags

A tag is a key-value pair that can be added to a file. This pair can then be used for searches (e.g. embedding searches) in order to narrow down the scope of the search. A file can have any number of tags. The following are reserved keys that cannot be used:

  • db_embedding_id
  • organization_id
  • user_id
  • organization_user_file_id

Carbon currently supports two data types for tag values - string and list<string>. Keys can only be string. If values other than string and list<string> are used, they're automatically converted to strings (e.g. 4 will become "4").

πŸ› οΈ Usage
result = carbon.files.create_user_file_tags(
  tags: {
        "key": "string_example",
    },
  organization_user_file_id: 1,
)
p result
βš™οΈ Parameters
tags: Hash<string, Tags1>
organization_user_file_id: Integer
πŸ”„ Return

UserFile

🌐 Endpoint

/create_user_file_tags POST

πŸ”™ Back to Table of Contents


carbon.files.delete_file_tags

Delete File Tags

πŸ› οΈ Usage
result = carbon.files.delete_file_tags(
  tags: [
        "string_example"
    ],
  organization_user_file_id: 1,
)
p result
βš™οΈ Parameters
tags: Array<String>
organization_user_file_id: Integer
πŸ”„ Return

UserFile

🌐 Endpoint

/delete_user_file_tags POST

πŸ”™ Back to Table of Contents


carbon.files.delete_many

Deprecated

Delete Files Endpoint

πŸ› οΈ Usage
result = carbon.files.delete_many(
  file_ids: [
        1
    ],
  sync_statuses: [
        "string_example"
    ],
  delete_non_synced_only: false,
  send_webhook: false,
  delete_child_files: false,
)
p result
βš™οΈ Parameters
file_ids: Array<Integer>
sync_statuses: Array<ExternalFileSyncStatuses>
delete_non_synced_only: Boolean
send_webhook: Boolean
delete_child_files: Boolean
πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/delete_files POST

πŸ”™ Back to Table of Contents


carbon.files.delete_v2

Delete Files V2 Endpoint

πŸ› οΈ Usage
result = carbon.files.delete_v2(
  filters: {
        "include_all_children" => false,
        "non_synced_only" => false,
    },
  send_webhook: false,
  preserve_file_record: false,
)
p result
βš™οΈ Parameters
filters: OrganizationUserFilesToSyncFilters
send_webhook: Boolean
preserve_file_record: Boolean

Whether or not to delete all data related to the file from the database, BUT to preserve the file metadata, allowing for resyncs. By default preserve_file_record is false, which means that all data related to the file as well as its metadata will be deleted. Note that even if preserve_file_record is true, raw files uploaded via the uploadfile endpoint still cannot be resynced.

πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/delete_files_v2 POST

πŸ”™ Back to Table of Contents


carbon.files.get_parsed_file

Deprecated

This route is deprecated. Use /user_files_v2 instead.

πŸ› οΈ Usage
result = carbon.files.get_parsed_file(
  file_id: 1,
)
p result
βš™οΈ Parameters
file_id: Integer
πŸ”„ Return

PresignedURLResponse

🌐 Endpoint

/parsed_file/{file_id} GET

πŸ”™ Back to Table of Contents


carbon.files.get_raw_file

Deprecated

This route is deprecated. Use /user_files_v2 instead.

πŸ› οΈ Usage
result = carbon.files.get_raw_file(
  file_id: 1,
)
p result
βš™οΈ Parameters
file_id: Integer
πŸ”„ Return

PresignedURLResponse

🌐 Endpoint

/raw_file/{file_id} GET

πŸ”™ Back to Table of Contents


carbon.files.modify_cold_storage_parameters

Modify Cold Storage Parameters

πŸ› οΈ Usage
result = carbon.files.modify_cold_storage_parameters(
  filters: {
        "include_all_children" => false,
        "non_synced_only" => false,
    },
  enable_cold_storage: true,
  hot_storage_time_to_live: 1,
)
p result
βš™οΈ Parameters
filters: OrganizationUserFilesToSyncFilters
enable_cold_storage: Boolean
hot_storage_time_to_live: Integer
🌐 Endpoint

/modify_cold_storage_parameters POST

πŸ”™ Back to Table of Contents


carbon.files.move_to_hot_storage

Move To Hot Storage

πŸ› οΈ Usage
result = carbon.files.move_to_hot_storage(
  filters: {
        "include_all_children" => false,
        "non_synced_only" => false,
    },
)
p result
βš™οΈ Parameters
filters: OrganizationUserFilesToSyncFilters
🌐 Endpoint

/move_to_hot_storage POST

πŸ”™ Back to Table of Contents


carbon.files.query_user_files

For pre-filtering documents, using tags_v2 is preferred to using tags (which is now deprecated). If both tags_v2 and tags are specified, tags is ignored. tags_v2 enables building complex filters through the use of "AND", "OR", and negation logic. Take the below input as an example:

{
    "OR": [
        {
            "key": "subject",
            "value": "holy-bible",
            "negate": false
        },
        {
            "key": "person-of-interest",
            "value": "jesus christ",
            "negate": false
        },
        {
            "key": "genre",
            "value": "religion",
            "negate": true
        }
        {
            "AND": [
                {
                    "key": "subject",
                    "value": "tao-te-ching",
                    "negate": false
                },
                {
                    "key": "author",
                    "value": "lao-tzu",
                    "negate": false
                }
            ]
        }
    ]
}

In this case, files will be filtered such that:

  1. "subject" = "holy-bible" OR
  2. "person-of-interest" = "jesus christ" OR
  3. "genre" != "religion" OR
  4. "subject" = "tao-te-ching" AND "author" = "lao-tzu"

Note that the top level of the query must be either an "OR" or "AND" array. Currently, nesting is limited to 3. For tag blocks (those with "key", "value", and "negate" keys), the following typing rules apply:

  1. "key" isn't optional and must be a string
  2. "value" isn't optional and can be any or list[any]
  3. "negate" is optional and must be true or false. If present and true, then the filter block is negated in the resulting query. It is false by default.
πŸ› οΈ Usage
result = carbon.files.query_user_files(
  pagination: {
        "limit" => 10,
        "offset" => 0,
        "starting_id" => 0,
    },
  order_by: "created_at",
  order_dir: "desc",
  filters: {
        "include_all_children" => false,
        "non_synced_only" => false,
    },
  include_raw_file: true,
  include_parsed_text_file: true,
  include_additional_files: true,
  presigned_url_expiry_time_seconds: 3600,
)
p result
βš™οΈ Parameters
pagination: Pagination
order_by: OrganizationUserFilesToSyncOrderByTypes
order_dir: OrderDir
filters: OrganizationUserFilesToSyncFilters
include_raw_file: Boolean

If true, the query will return presigned URLs for the raw file. Only relevant for the /user_files_v2 endpoint.

include_parsed_text_file: Boolean

If true, the query will return presigned URLs for the parsed text file. Only relevant for the /user_files_v2 endpoint.

include_additional_files: Boolean

If true, the query will return presigned URLs for additional files. Only relevant for the /user_files_v2 endpoint.

presigned_url_expiry_time_seconds: Integer

The expiry time for the presigned URLs. Only relevant for the /user_files_v2 endpoint.

πŸ”„ Return

UserFilesV2

🌐 Endpoint

/user_files_v2 POST

πŸ”™ Back to Table of Contents


carbon.files.query_user_files_deprecated

Deprecated

This route is deprecated. Use /user_files_v2 instead.

πŸ› οΈ Usage
result = carbon.files.query_user_files_deprecated(
  pagination: {
        "limit" => 10,
        "offset" => 0,
        "starting_id" => 0,
    },
  order_by: "created_at",
  order_dir: "desc",
  filters: {
        "include_all_children" => false,
        "non_synced_only" => false,
    },
  include_raw_file: true,
  include_parsed_text_file: true,
  include_additional_files: true,
  presigned_url_expiry_time_seconds: 3600,
)
p result
βš™οΈ Parameters
pagination: Pagination
order_by: OrganizationUserFilesToSyncOrderByTypes
order_dir: OrderDir
filters: OrganizationUserFilesToSyncFilters
include_raw_file: Boolean

If true, the query will return presigned URLs for the raw file. Only relevant for the /user_files_v2 endpoint.

include_parsed_text_file: Boolean

If true, the query will return presigned URLs for the parsed text file. Only relevant for the /user_files_v2 endpoint.

include_additional_files: Boolean

If true, the query will return presigned URLs for additional files. Only relevant for the /user_files_v2 endpoint.

presigned_url_expiry_time_seconds: Integer

The expiry time for the presigned URLs. Only relevant for the /user_files_v2 endpoint.

πŸ”„ Return

UserFile

🌐 Endpoint

/user_files POST

πŸ”™ Back to Table of Contents


carbon.files.resync

Resync File

πŸ› οΈ Usage
result = carbon.files.resync(
  file_id: 1,
  chunk_size: 1,
  chunk_overlap: 1,
  force_embedding_generation: false,
  skip_file_processing: false,
)
p result
βš™οΈ Parameters
file_id: Integer
chunk_size: Integer
chunk_overlap: Integer
force_embedding_generation: Boolean
skip_file_processing: Boolean
πŸ”„ Return

UserFile

🌐 Endpoint

/resync_file POST

πŸ”™ Back to Table of Contents


carbon.files.upload

This endpoint is used to directly upload local files to Carbon. The POST request should be a multipart form request. Note that the set_page_as_boundary query parameter is applicable only to PDFs for now. When this value is set, PDF chunks are at most one page long. Additional information can be retrieved for each chunk, however, namely the coordinates of the bounding box around the chunk (this can be used for things like text highlighting). Following is a description of all possible query parameters:

  • chunk_size: the chunk size (in tokens) applied when splitting the document
  • chunk_overlap: the chunk overlap (in tokens) applied when splitting the document
  • skip_embedding_generation: whether or not to skip the generation of chunks and embeddings
  • set_page_as_boundary: described above
  • embedding_model: the model used to generate embeddings for the document chunks
  • use_ocr: whether or not to use OCR as a preprocessing step prior to generating chunks. Valid for PDFs, JPEGs, and PNGs
  • generate_sparse_vectors: whether or not to generate sparse vectors for the file. Required for hybrid search.
  • prepend_filename_to_chunks: whether or not to prepend the filename to the chunk text

Carbon supports multiple models for use in generating embeddings for files. For images, we support Vertex AI's multimodal model; for text, we support OpenAI's text-embedding-ada-002 and Cohere's embed-multilingual-v3.0. The model can be specified via the embedding_model parameter (in the POST body for /embeddings, and a query parameter in /uploadfile). If no model is supplied, the text-embedding-ada-002 is used by default. When performing embedding queries, embeddings from files that used the specified model will be considered in the query. For example, if files A and B have embeddings generated with OPENAI, and files C and D have embeddings generated with COHERE_MULTILINGUAL_V3, then by default, queries will only consider files A and B. If COHERE_MULTILINGUAL_V3 is specified as the embedding_model in /embeddings, then only files C and D will be considered. Make sure that the set of all files you want considered for a query have embeddings generated via the same model. For now, do not set VERTEX_MULTIMODAL as an embedding_model. This model is used automatically by Carbon when it detects an image file.

πŸ› οΈ Usage
result = carbon.files.upload(
  file: File.open("path/to/file", "rb"),
  chunk_size: 1,
  chunk_overlap: 1,
  skip_embedding_generation: false,
  set_page_as_boundary: false,
  embedding_model: "string_example",
  use_ocr: false,
  generate_sparse_vectors: false,
  prepend_filename_to_chunks: false,
  max_items_per_chunk: 1,
  parse_pdf_tables_with_ocr: false,
  detect_audio_language: false,
  transcription_service: "assemblyai",
  include_speaker_labels: false,
  media_type: "TEXT",
  split_rows: false,
  enable_cold_storage: false,
  hot_storage_time_to_live: 1,
  generate_chunks_only: false,
  store_file_only: false,
)
p result
βš™οΈ Parameters
file: File
chunk_size: Integer

Chunk size in tiktoken tokens to be used when processing file.

chunk_overlap: Integer

Chunk overlap in tiktoken tokens to be used when processing file.

skip_embedding_generation: Boolean

Flag to control whether or not embeddings should be generated and stored when processing file.

set_page_as_boundary: Boolean

Flag to control whether or not to set the a page's worth of content as the maximum amount of content that can appear in a chunk. Only valid for PDFs. See description route description for more information.

embedding_model: EmbeddingModel

Embedding model that will be used to embed file chunks.

use_ocr: Boolean

Whether or not to use OCR when processing files. Valid for PDFs, JPEGs, and PNGs. Useful for documents with tables, images, and/or scanned text.

generate_sparse_vectors: Boolean

Whether or not to generate sparse vectors for the file. This is required for the file to be a candidate for hybrid search.

prepend_filename_to_chunks: Boolean

Whether or not to prepend the file's name to chunks.

max_items_per_chunk: Integer

Number of objects per chunk. For csv, tsv, xlsx, and json files only.

parse_pdf_tables_with_ocr: Boolean

Whether to use rich table parsing when use_ocr is enabled.

detect_audio_language: Boolean

Whether to automatically detect the language of the uploaded audio file.

transcription_service: TranscriptionServiceNullable

The transcription service to use for audio files. If no service is specified, 'deepgram' will be used.

include_speaker_labels: Boolean

Detect multiple speakers and label segments of speech by speaker for audio files.

media_type: FileContentTypesNullable

The media type of the file. If not provided, it will be inferred from the file extension.

split_rows: Boolean

Whether to split tabular rows into chunks. Currently only valid for CSV, TSV, and XLSX files.

enable_cold_storage: Boolean

Enable cold storage for the file. If set to true, the file will be moved to cold storage after a certain period of inactivity. Default is false.

hot_storage_time_to_live: Integer

Time in days after which the file will be moved to cold storage. Must be one of [1, 3, 7, 14, 30].

generate_chunks_only: Boolean

If this flag is enabled, the file will be chunked and stored with Carbon, but no embeddings will be generated. This overrides the skip_embedding_generation flag.

store_file_only: Boolean

If this flag is enabled, the file will be stored with Carbon, but no processing will be done.

πŸ”„ Return

UserFile

🌐 Endpoint

/uploadfile POST

πŸ”™ Back to Table of Contents


carbon.files.upload_from_url

Create Upload File From Url

πŸ› οΈ Usage
result = carbon.files.upload_from_url(
  url: "string_example",
  file_name: "string_example",
  chunk_size: 1,
  chunk_overlap: 1,
  skip_embedding_generation: false,
  set_page_as_boundary: false,
  embedding_model: "OPENAI",
  generate_sparse_vectors: false,
  use_textract: false,
  prepend_filename_to_chunks: false,
  max_items_per_chunk: 1,
  parse_pdf_tables_with_ocr: false,
  detect_audio_language: false,
  transcription_service: "assemblyai",
  include_speaker_labels: false,
  media_type: "TEXT",
  split_rows: false,
  cold_storage_params: {
        "enable_cold_storage" => false,
    },
  generate_chunks_only: false,
  store_file_only: false,
)
p result
βš™οΈ Parameters
url: String
file_name: String
chunk_size: Integer
chunk_overlap: Integer
skip_embedding_generation: Boolean
set_page_as_boundary: Boolean
embedding_model: EmbeddingGenerators
generate_sparse_vectors: Boolean
use_textract: Boolean
prepend_filename_to_chunks: Boolean
max_items_per_chunk: Integer

Number of objects per chunk. For csv, tsv, xlsx, and json files only.

parse_pdf_tables_with_ocr: Boolean
detect_audio_language: Boolean
transcription_service: TranscriptionServiceNullable
include_speaker_labels: Boolean
media_type: FileContentTypesNullable
split_rows: Boolean
cold_storage_params: ColdStorageProps
generate_chunks_only: Boolean

If this flag is enabled, the file will be chunked and stored with Carbon, but no embeddings will be generated. This overrides the skip_embedding_generation flag.

store_file_only: Boolean

If this flag is enabled, the file will be stored with Carbon, but no processing will be done.

πŸ”„ Return

UserFile

🌐 Endpoint

/upload_file_from_url POST

πŸ”™ Back to Table of Contents


carbon.files.upload_text

Carbon supports multiple models for use in generating embeddings for files. For images, we support Vertex AI's multimodal model; for text, we support OpenAI's text-embedding-ada-002 and Cohere's embed-multilingual-v3.0. The model can be specified via the embedding_model parameter (in the POST body for /embeddings, and a query parameter in /uploadfile). If no model is supplied, the text-embedding-ada-002 is used by default. When performing embedding queries, embeddings from files that used the specified model will be considered in the query. For example, if files A and B have embeddings generated with OPENAI, and files C and D have embeddings generated with COHERE_MULTILINGUAL_V3, then by default, queries will only consider files A and B. If COHERE_MULTILINGUAL_V3 is specified as the embedding_model in /embeddings, then only files C and D will be considered. Make sure that the set of all files you want considered for a query have embeddings generated via the same model. For now, do not set VERTEX_MULTIMODAL as an embedding_model. This model is used automatically by Carbon when it detects an image file.

πŸ› οΈ Usage
result = carbon.files.upload_text(
  contents: "aaaaa",
  name: "string_example",
  chunk_size: 1,
  chunk_overlap: 1,
  skip_embedding_generation: false,
  overwrite_file_id: 1,
  embedding_model: "OPENAI",
  generate_sparse_vectors: false,
  cold_storage_params: {
        "enable_cold_storage" => false,
    },
  generate_chunks_only: false,
  store_file_only: false,
)
p result
βš™οΈ Parameters
contents: String
name: String
chunk_size: Integer
chunk_overlap: Integer
skip_embedding_generation: Boolean
overwrite_file_id: Integer
embedding_model: EmbeddingGeneratorsNullable
generate_sparse_vectors: Boolean
cold_storage_params: ColdStorageProps
generate_chunks_only: Boolean

If this flag is enabled, the file will be chunked and stored with Carbon, but no embeddings will be generated. This overrides the skip_embedding_generation flag.

store_file_only: Boolean

If this flag is enabled, the file will be stored with Carbon, but no processing will be done.

πŸ”„ Return

UserFile

🌐 Endpoint

/upload_text POST

πŸ”™ Back to Table of Contents


carbon.github.get_issue

Issue

πŸ› οΈ Usage
result = carbon.github.get_issue(
  issue_number: 1,
  include_remote_data: false,
  data_source_id: 1,
  repository: "string_example",
)
p result
βš™οΈ Parameters
issue_number: Integer
include_remote_data: Boolean
data_source_id: Integer
repository: String
πŸ”„ Return

Issue

🌐 Endpoint

/integrations/data/github/issues/{issue_number} GET

πŸ”™ Back to Table of Contents


carbon.github.get_issues

Issues

πŸ› οΈ Usage
result = carbon.github.get_issues(
  data_source_id: 1,
  repository: "string_example",
  include_remote_data: false,
  page: 1,
  page_size: 30,
  next_cursor: "string_example",
  filters: {
        "state" => "closed",
    },
  order_by: "created",
  order_dir: "asc",
)
p result
βš™οΈ Parameters
data_source_id: Integer
repository: String

Full name of the repository, denoted as {owner}/{repo}

include_remote_data: Boolean
page: Integer
page_size: Integer
next_cursor: String
filters: IssuesFilter
order_by: IssuesOrderBy
order_dir: OrderDirV2Nullable
πŸ”„ Return

IssuesResponse

🌐 Endpoint

/integrations/data/github/issues POST

πŸ”™ Back to Table of Contents


carbon.github.get_pr

Get Pr

πŸ› οΈ Usage
result = carbon.github.get_pr(
  pull_number: 1,
  include_remote_data: false,
  data_source_id: 1,
  repository: "string_example",
)
p result
βš™οΈ Parameters
pull_number: Integer
include_remote_data: Boolean
data_source_id: Integer
repository: String
πŸ”„ Return

PullRequestExtended

🌐 Endpoint

/integrations/data/github/pull_requests/{pull_number} GET

πŸ”™ Back to Table of Contents


carbon.github.get_pr_comments

Pr Comments

πŸ› οΈ Usage
result = carbon.github.get_pr_comments(
  data_source_id: 1,
  repository: "string_example",
  pull_number: 1,
  include_remote_data: false,
  page: 1,
  page_size: 30,
  next_cursor: "string_example",
  order_by: "created",
  order_dir: "asc",
)
p result
βš™οΈ Parameters
data_source_id: Integer
repository: String

Full name of the repository, denoted as {owner}/{repo}

pull_number: Integer
include_remote_data: Boolean
page: Integer
page_size: Integer
next_cursor: String
order_by: CommentsOrderBy
order_dir: OrderDirV2Nullable
πŸ”„ Return

CommentsResponse

🌐 Endpoint

/integrations/data/github/pull_requests/comments POST

πŸ”™ Back to Table of Contents


carbon.github.get_pr_commits

Pr Commits

πŸ› οΈ Usage
result = carbon.github.get_pr_commits(
  data_source_id: 1,
  repository: "string_example",
  pull_number: 1,
  include_remote_data: false,
  page: 1,
  page_size: 30,
  next_cursor: "string_example",
)
p result
βš™οΈ Parameters
data_source_id: Integer
repository: String

Full name of the repository, denoted as {owner}/{repo}

pull_number: Integer
include_remote_data: Boolean
page: Integer
page_size: Integer
next_cursor: String
πŸ”„ Return

CommitsResponse

🌐 Endpoint

/integrations/data/github/pull_requests/commits POST

πŸ”™ Back to Table of Contents


carbon.github.get_pr_files

Pr Files

πŸ› οΈ Usage
result = carbon.github.get_pr_files(
  data_source_id: 1,
  repository: "string_example",
  pull_number: 1,
  include_remote_data: false,
  page: 1,
  page_size: 30,
  next_cursor: "string_example",
)
p result
βš™οΈ Parameters
data_source_id: Integer
repository: String

Full name of the repository, denoted as {owner}/{repo}

pull_number: Integer
include_remote_data: Boolean
page: Integer
page_size: Integer
next_cursor: String
πŸ”„ Return

FilesResponse

🌐 Endpoint

/integrations/data/github/pull_requests/files POST

πŸ”™ Back to Table of Contents


carbon.github.get_pull_requests

Get Prs

πŸ› οΈ Usage
result = carbon.github.get_pull_requests(
  data_source_id: 1,
  repository: "string_example",
  include_remote_data: false,
  page: 1,
  page_size: 30,
  next_cursor: "string_example",
  filters: {
        "state" => "closed",
    },
  order_by: "created",
  order_dir: "asc",
)
p result
βš™οΈ Parameters
data_source_id: Integer
repository: String

Full name of the repository, denoted as {owner}/{repo}

include_remote_data: Boolean
page: Integer
page_size: Integer
next_cursor: String
filters: PullRequestFilters
order_by: PROrderBy
order_dir: OrderDirV2Nullable
πŸ”„ Return

PullRequestResponse

🌐 Endpoint

/integrations/data/github/pull_requests POST

πŸ”™ Back to Table of Contents


carbon.integrations.cancel

Cancel Data Source Items Sync

πŸ› οΈ Usage
result = carbon.integrations.cancel(
  data_source_id: 1,
)
p result
βš™οΈ Parameters
data_source_id: Integer
πŸ”„ Return

OrganizationUserDataSourceAPI

🌐 Endpoint

/integrations/items/sync/cancel POST

πŸ”™ Back to Table of Contents


carbon.integrations.connect_data_source

Connect Data Source

πŸ› οΈ Usage
result = carbon.integrations.connect_data_source(
  authentication: {
        "source" => "GOOGLE_DRIVE",
        "access_token" => "access_token_example",
    },
  sync_options: {
        "chunk_size" => 1500,
        "chunk_overlap" => 20,
        "skip_embedding_generation" => false,
        "embedding_model" => "OPENAI",
        "generate_sparse_vectors" => false,
        "prepend_filename_to_chunks" => false,
        "sync_files_on_connection" => true,
        "set_page_as_boundary" => false,
        "enable_file_picker" => true,
        "sync_source_items" => true,
        "incremental_sync" => false,
    },
)
p result
βš™οΈ Parameters
authentication: AuthenticationProperty
sync_options: SyncOptions
πŸ”„ Return

ConnectDataSourceResponse

🌐 Endpoint

/integrations/connect POST

πŸ”™ Back to Table of Contents


carbon.integrations.connect_document360

You will need an access token to connect your Document360 account. To obtain an access token, follow the steps highlighted here https://apidocs.document360.com/apidocs/api-token.

πŸ› οΈ Usage
result = carbon.integrations.connect_document360(
  account_email: "string_example",
  access_token: "string_example",
  tags: {},
  chunk_size: 1500,
  chunk_overlap: 20,
  skip_embedding_generation: false,
  embedding_model: "OPENAI",
  generate_sparse_vectors: false,
  prepend_filename_to_chunks: false,
  sync_files_on_connection: true,
  request_id: "string_example",
  sync_source_items: true,
  file_sync_config: {
        "auto_synced_source_types" => ["ARTICLE"],
        "sync_attachments" => false,
        "detect_audio_language" => false,
        "transcription_service" => "assemblyai",
        "include_speaker_labels" => false,
        "split_rows" => false,
        "generate_chunks_only" => false,
        "store_file_only" => false,
        "skip_file_processing" => false,
        "parsed_text_format" => "PLAIN_TEXT",
    },
  data_source_tags: {},
)
p result
βš™οΈ Parameters
account_email: String

This email will be used to identify your carbon data source. It should have access to the Document360 account you wish to connect.

access_token: String
tags: Object
chunk_size: Integer
chunk_overlap: Integer
skip_embedding_generation: Boolean
embedding_model: EmbeddingGenerators
generate_sparse_vectors: Boolean
prepend_filename_to_chunks: Boolean
sync_files_on_connection: Boolean
request_id: String
sync_source_items: Boolean

Enabling this flag will fetch all available content from the source to be listed via list items endpoint

file_sync_config: FileSyncConfigNullable
data_source_tags: Object

Tags to be associated with the data source. If the data source already has tags set, then an upsert will be performed.

πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/integrations/document360 POST

πŸ”™ Back to Table of Contents


carbon.integrations.connect_freshdesk

Refer this article to obtain an API key https://support.freshdesk.com/en/support/solutions/articles/215517. Make sure that your API key has the permission to read solutions from your account and you are on a paid plan. Once you have an API key, you can make a request to this endpoint along with your freshdesk domain. This will trigger an automatic sync of the articles in your "solutions" tab. Additional parameters below can be used to associate data with the synced articles or modify the sync behavior.

πŸ› οΈ Usage
result = carbon.integrations.connect_freshdesk(
  domain: "string_example",
  api_key: "string_example",
  tags: {},
  chunk_size: 1500,
  chunk_overlap: 20,
  skip_embedding_generation: false,
  embedding_model: "OPENAI",
  generate_sparse_vectors: false,
  prepend_filename_to_chunks: false,
  sync_files_on_connection: true,
  request_id: "string_example",
  sync_source_items: true,
  file_sync_config: {
        "auto_synced_source_types" => ["ARTICLE"],
        "sync_attachments" => false,
        "detect_audio_language" => false,
        "transcription_service" => "assemblyai",
        "include_speaker_labels" => false,
        "split_rows" => false,
        "generate_chunks_only" => false,
        "store_file_only" => false,
        "skip_file_processing" => false,
        "parsed_text_format" => "PLAIN_TEXT",
    },
  data_source_tags: {},
)
p result
βš™οΈ Parameters
domain: String
api_key: String
tags: Object
chunk_size: Integer
chunk_overlap: Integer
skip_embedding_generation: Boolean
embedding_model: EmbeddingGeneratorsNullable
generate_sparse_vectors: Boolean
prepend_filename_to_chunks: Boolean
sync_files_on_connection: Boolean
request_id: String
sync_source_items: Boolean

Enabling this flag will fetch all available content from the source to be listed via list items endpoint

file_sync_config: FileSyncConfigNullable
data_source_tags: Object

Tags to be associated with the data source. If the data source already has tags set, then an upsert will be performed.

πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/integrations/freshdesk POST

πŸ”™ Back to Table of Contents


carbon.integrations.connect_gitbook

You will need an access token to connect your Gitbook account. Note that the permissions will be defined by the user generating access token so make sure you have the permission to access spaces you will be syncing. Refer this article for more details https://developer.gitbook.com/gitbook-api/authentication. Additionally, you need to specify the name of organization you will be syncing data from.

πŸ› οΈ Usage
result = carbon.integrations.connect_gitbook(
  organization: "string_example",
  access_token: "string_example",
  tags: {},
  chunk_size: 1500,
  chunk_overlap: 20,
  skip_embedding_generation: false,
  embedding_model: "OPENAI",
  generate_sparse_vectors: false,
  prepend_filename_to_chunks: false,
  sync_files_on_connection: true,
  request_id: "string_example",
  sync_source_items: true,
  file_sync_config: {
        "auto_synced_source_types" => ["ARTICLE"],
        "sync_attachments" => false,
        "detect_audio_language" => false,
        "transcription_service" => "assemblyai",
        "include_speaker_labels" => false,
        "split_rows" => false,
        "generate_chunks_only" => false,
        "store_file_only" => false,
        "skip_file_processing" => false,
        "parsed_text_format" => "PLAIN_TEXT",
    },
  data_source_tags: {},
)
p result
βš™οΈ Parameters
organization: String
access_token: String
tags: Object
chunk_size: Integer
chunk_overlap: Integer
skip_embedding_generation: Boolean
embedding_model: EmbeddingGenerators
generate_sparse_vectors: Boolean
prepend_filename_to_chunks: Boolean
sync_files_on_connection: Boolean
request_id: String
sync_source_items: Boolean

Enabling this flag will fetch all available content from the source to be listed via list items endpoint

file_sync_config: FileSyncConfigNullable
data_source_tags: Object

Tags to be associated with the data source. If the data source already has tags set, then an upsert will be performed.

πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/integrations/gitbook POST

πŸ”™ Back to Table of Contents


carbon.integrations.connect_guru

You will need an access token to connect your Guru account. To obtain an access token, follow the steps highlighted here https://help.getguru.com/docs/gurus-api#obtaining-a-user-token. The username should be your Guru username.

πŸ› οΈ Usage
result = carbon.integrations.connect_guru(
  username: "string_example",
  access_token: "string_example",
  tags: {},
  chunk_size: 1500,
  chunk_overlap: 20,
  skip_embedding_generation: false,
  embedding_model: "OPENAI",
  generate_sparse_vectors: false,
  prepend_filename_to_chunks: false,
  sync_files_on_connection: true,
  request_id: "string_example",
  sync_source_items: true,
  file_sync_config: {
        "auto_synced_source_types" => ["ARTICLE"],
        "sync_attachments" => false,
        "detect_audio_language" => false,
        "transcription_service" => "assemblyai",
        "include_speaker_labels" => false,
        "split_rows" => false,
        "generate_chunks_only" => false,
        "store_file_only" => false,
        "skip_file_processing" => false,
        "parsed_text_format" => "PLAIN_TEXT",
    },
  data_source_tags: {},
)
p result
βš™οΈ Parameters
username: String
access_token: String
tags: Object
chunk_size: Integer
chunk_overlap: Integer
skip_embedding_generation: Boolean
embedding_model: EmbeddingGenerators
generate_sparse_vectors: Boolean
prepend_filename_to_chunks: Boolean
sync_files_on_connection: Boolean
request_id: String
sync_source_items: Boolean

Enabling this flag will fetch all available content from the source to be listed via list items endpoint

file_sync_config: FileSyncConfigNullable
data_source_tags: Object

Tags to be associated with the data source. If the data source already has tags set, then an upsert will be performed.

πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/integrations/guru POST

πŸ”™ Back to Table of Contents


carbon.integrations.create_aws_iam_user

This endpoint can be used to connect S3 as well as Digital Ocean Spaces (S3 compatible)
For S3, create a new IAM user with permissions to:

  1. List all buckets.
  2. Read from the specific buckets and objects to sync with Carbon. Ensure any future buckets or objects carry the same permissions.
Once created, generate an access key for this user and share the credentials with us. We recommend testing this key beforehand. For Digital Ocean Spaces, generate the above credentials in your Applications and API page here https://cloud.digitalocean.com/account/api/spaces. Endpoint URL is required to connect Digital Ocean Spaces. It should look like <>.digitaloceanspaces.com
πŸ› οΈ Usage
result = carbon.integrations.create_aws_iam_user(
  access_key: "string_example",
  access_key_secret: "string_example",
  sync_source_items: true,
  endpoint_url: "string_example",
  data_source_tags: {},
)
p result
βš™οΈ Parameters
access_key: String
access_key_secret: String
sync_source_items: Boolean

Enabling this flag will fetch all available content from the source to be listed via list items endpoint

endpoint_url: String

You can specify a Digital Ocean endpoint URL to connect a Digital Ocean Space through this endpoint. The URL should be of format .digitaloceanspaces.com. It's not required for S3 buckets.

data_source_tags: Object

Tags to be associated with the data source. If the data source already has tags set, then an upsert will be performed.

πŸ”„ Return

OrganizationUserDataSourceAPI

🌐 Endpoint

/integrations/s3 POST

πŸ”™ Back to Table of Contents


carbon.integrations.get_oauth_url

This endpoint can be used to generate the following URLs

  • An OAuth URL for OAuth based connectors
  • A file syncing URL which skips the OAuth flow if the user already has a valid access token and takes them to the success state.
πŸ› οΈ Usage
result = carbon.integrations.get_oauth_url(
  service: "BOX",
  tags: None,
  scope: "string_example",
  scopes: [],
  chunk_size: 1500,
  chunk_overlap: 20,
  skip_embedding_generation: false,
  embedding_model: "OPENAI",
  zendesk_subdomain: "string_example",
  microsoft_tenant: "string_example",
  sharepoint_site_name: "string_example",
  confluence_subdomain: "string_example",
  generate_sparse_vectors: false,
  prepend_filename_to_chunks: false,
  max_items_per_chunk: 1,
  salesforce_domain: "string_example",
  sync_files_on_connection: true,
  set_page_as_boundary: false,
  data_source_id: 1,
  connecting_new_account: false,
  request_id: "string_example",
  use_ocr: false,
  parse_pdf_tables_with_ocr: false,
  enable_file_picker: true,
  sync_source_items: true,
  incremental_sync: false,
  file_sync_config: {
        "auto_synced_source_types" => ["ARTICLE"],
        "sync_attachments" => false,
        "detect_audio_language" => false,
        "transcription_service" => "assemblyai",
        "include_speaker_labels" => false,
        "split_rows" => false,
        "generate_chunks_only" => false,
        "store_file_only" => false,
        "skip_file_processing" => false,
        "parsed_text_format" => "PLAIN_TEXT",
    },
  automatically_open_file_picker: true,
  gong_account_email: "string_example",
  servicenow_credentials: {
        "instance_subdomain" => "instance_subdomain_example",
        "client_id" => "client_id_example",
        "client_secret" => "client_secret_example",
        "redirect_uri" => "redirect_uri_example",
    },
  data_source_tags: {},
)
p result
βš™οΈ Parameters
service: OauthBasedConnectors
tags: Object
scope: String
scopes: Array<String>

List of scopes to request from the OAuth provider. Please that the scopes will be used as it is, not combined with the default props that Carbon uses. One scope should be one array element.

chunk_size: Integer
chunk_overlap: Integer
skip_embedding_generation: Boolean
embedding_model: EmbeddingGeneratorsNullable
zendesk_subdomain: String
microsoft_tenant: String
sharepoint_site_name: String
confluence_subdomain: String
generate_sparse_vectors: Boolean
prepend_filename_to_chunks: Boolean
max_items_per_chunk: Integer

Number of objects per chunk. For csv, tsv, xlsx, and json files only.

salesforce_domain: String
sync_files_on_connection: Boolean

Used to specify whether Carbon should attempt to sync all your files automatically when authorization is complete. This is only supported for a subset of connectors and will be ignored for the rest. Supported connectors: Intercom, Zendesk, Gitbook, Confluence, Salesforce, Freshdesk

set_page_as_boundary: Boolean
data_source_id: Integer

Used to specify a data source to sync from if you have multiple connected. It can be skipped if you only have one data source of that type connected or are connecting a new account.

connecting_new_account: Boolean

Used to connect a new data source. If not specified, we will attempt to create a sync URL for an existing data source based on type and ID.

request_id: String

This request id will be added to all files that get synced using the generated OAuth URL

use_ocr: Boolean

Enable OCR for files that support it. Supported formats: pdf, png, jpg

parse_pdf_tables_with_ocr: Boolean
enable_file_picker: Boolean

Enable integration's file picker for sources that support it. Supported sources: BOX, DROPBOX, GOOGLE_DRIVE, ONEDRIVE, SHAREPOINT

sync_source_items: Boolean

Enabling this flag will fetch all available content from the source to be listed via list items endpoint

incremental_sync: Boolean

Only sync files if they have not already been synced or if the embedding properties have changed. This flag is currently supported by ONEDRIVE, GOOGLE_DRIVE, BOX, DROPBOX, INTERCOM, GMAIL, OUTLOOK, ZENDESK, CONFLUENCE, NOTION, SHAREPOINT, SERVICENOW. It will be ignored for other data sources.

file_sync_config: FileSyncConfigNullable
automatically_open_file_picker: Boolean

Automatically open source file picker after the OAuth flow is complete. This flag is currently supported by BOX, DROPBOX, GOOGLE_DRIVE, ONEDRIVE, SHAREPOINT. It will be ignored for other data sources.

gong_account_email: String

If you are connecting a Gong account, you need to input the email of the account you wish to connect. This email will be used to identify your carbon data source.

servicenow_credentials: ServiceNowCredentialsNullable
data_source_tags: Object

Tags to be associated with the data source. If the data source already has tags set, then an upsert will be performed.

πŸ”„ Return

OuthURLResponse

🌐 Endpoint

/integrations/oauth_url POST

πŸ”™ Back to Table of Contents


carbon.integrations.list_confluence_pages

Deprecated

This endpoint has been deprecated. Use /integrations/items/list instead.

To begin listing a user's Confluence pages, at least a data_source_id of a connected Confluence account must be specified. This base request returns a list of root pages for every space the user has access to in a Confluence instance. To traverse further down the user's page directory, additional requests to this endpoint can be made with the same data_source_id and with parent_id set to the id of page from a previous request. For convenience, the has_children property in each directory item in the response list will flag which pages will return non-empty lists of pages when set as the parent_id.

πŸ› οΈ Usage
result = carbon.integrations.list_confluence_pages(
  data_source_id: 1,
  parent_id: "string_example",
)
p result
βš™οΈ Parameters
data_source_id: Integer
parent_id: String
πŸ”„ Return

ListResponse

🌐 Endpoint

/integrations/confluence/list POST

πŸ”™ Back to Table of Contents


carbon.integrations.list_conversations

List all of your public and private channels, DMs, and Group DMs. The ID from response can be used as a filter to sync messages to Carbon
types: Comma separated list of types. Available types are im (DMs), mpim (group DMs), public_channel, and private_channel. Defaults to public_channel.
cursor: Used for pagination. If next_cursor is returned in response, you need to pass it as the cursor in the next request
data_source_id: Data source needs to be specified if you have linked multiple slack accounts
exclude_archived: Should archived conversations be excluded, defaults to true

πŸ› οΈ Usage
result = carbon.integrations.list_conversations(
  types: "public_channel",
  cursor: "string_example",
  data_source_id: 1,
  exclude_archived: true,
)
p result
βš™οΈ Parameters
types: String
cursor: String
data_source_id: Integer
exclude_archived: Boolean
🌐 Endpoint

/integrations/slack/conversations GET

πŸ”™ Back to Table of Contents


carbon.integrations.list_data_source_items

List Data Source Items

πŸ› οΈ Usage
result = carbon.integrations.list_data_source_items(
  data_source_id: 1,
  parent_id: "string_example",
  filters: {
    },
  pagination: {
        "limit" => 10,
        "offset" => 0,
        "starting_id" => 0,
    },
  order_by: "name",
  order_dir: "asc",
)
p result
βš™οΈ Parameters
data_source_id: Integer
parent_id: String
filters: ListItemsFiltersNullable
pagination: Pagination
order_by: ExternalSourceItemsOrderBy
order_dir: OrderDirV2
πŸ”„ Return

ListDataSourceItemsResponse

🌐 Endpoint

/integrations/items/list POST

πŸ”™ Back to Table of Contents


carbon.integrations.list_folders

After connecting your Outlook account, you can use this endpoint to list all of your folders on outlook. This includes both system folders like "inbox" and user created folders.

πŸ› οΈ Usage
result = carbon.integrations.list_folders(
  data_source_id: 1,
)
p result
βš™οΈ Parameters
data_source_id: Integer
🌐 Endpoint

/integrations/outlook/user_folders GET

πŸ”™ Back to Table of Contents


carbon.integrations.list_gitbook_spaces

After connecting your Gitbook account, you can use this endpoint to list all of your spaces under current organization.

πŸ› οΈ Usage
result = carbon.integrations.list_gitbook_spaces(
  data_source_id: 1,
)
p result
βš™οΈ Parameters
data_source_id: Integer
🌐 Endpoint

/integrations/gitbook/spaces GET

πŸ”™ Back to Table of Contents


carbon.integrations.list_labels

After connecting your Gmail account, you can use this endpoint to list all of your labels. User created labels will have the type "user" and Gmail's default labels will have the type "system"

πŸ› οΈ Usage
result = carbon.integrations.list_labels(
  data_source_id: 1,
)
p result
βš™οΈ Parameters
data_source_id: Integer
🌐 Endpoint

/integrations/gmail/user_labels GET

πŸ”™ Back to Table of Contents


carbon.integrations.list_outlook_categories

After connecting your Outlook account, you can use this endpoint to list all of your categories on outlook. We currently support listing up to 250 categories.

πŸ› οΈ Usage
result = carbon.integrations.list_outlook_categories(
  data_source_id: 1,
)
p result
βš™οΈ Parameters
data_source_id: Integer
🌐 Endpoint

/integrations/outlook/user_categories GET

πŸ”™ Back to Table of Contents


carbon.integrations.list_repos

Once you have connected your GitHub account, you can use this endpoint to list the repositories your account has access to. You can use a data source ID or username to fetch from a specific account.

πŸ› οΈ Usage
result = carbon.integrations.list_repos(
  per_page: 30,
  page: 1,
  data_source_id: 1,
)
p result
βš™οΈ Parameters
per_page: Integer
page: Integer
data_source_id: Integer
🌐 Endpoint

/integrations/github/repos GET

πŸ”™ Back to Table of Contents


carbon.integrations.list_sharepoint_sites

List all Sharepoint sites in the connected tenant. The site names from the response can be used as the site name when connecting a Sharepoint site. If site name is null in the response, then site name should be left null when connecting to the site.

This endpoint requires an additional Sharepoint scope: "Sites.Read.All". Include this scope along with the default Sharepoint scopes to list Sharepoint sites, connect to a site, and finally sync files from the site. The default Sharepoint scopes are: [o, p, e, n, i, d, , o, f, f, l, i, n, e, _, a, c, c, e, s, s, , U, s, e, r, ., R, e, a, d, , F, i, l, e, s, ., R, e, a, d, ., A, l, l].

data_soure_id: Data source needs to be specified if you have linked multiple Sharepoint accounts cursor: Used for pagination. If next_cursor is returned in response, you need to pass it as the cursor in the next request

πŸ› οΈ Usage
result = carbon.integrations.list_sharepoint_sites(
  data_source_id: 1,
  cursor: "string_example",
)
p result
βš™οΈ Parameters
data_source_id: Integer
cursor: String
🌐 Endpoint

/integrations/sharepoint/sites/list GET

πŸ”™ Back to Table of Contents


carbon.integrations.sync_azure_blob_files

After optionally loading the items via /integrations/items/sync and integrations/items/list, use the container name and file name as the ID in this endpoint to sync them into Carbon. Additional parameters below can associate data with the selected items or modify the sync behavior

πŸ› οΈ Usage
result = carbon.integrations.sync_azure_blob_files(
  ids: [
        {
        }
    ],
  tags: {},
  chunk_size: 1500,
  chunk_overlap: 20,
  skip_embedding_generation: false,
  embedding_model: "OPENAI",
  generate_sparse_vectors: false,
  prepend_filename_to_chunks: false,
  max_items_per_chunk: 1,
  set_page_as_boundary: false,
  data_source_id: 1,
  request_id: "string_example",
  use_ocr: false,
  parse_pdf_tables_with_ocr: false,
  file_sync_config: {
        "auto_synced_source_types" => ["ARTICLE"],
        "sync_attachments" => false,
        "detect_audio_language" => false,
        "transcription_service" => "assemblyai",
        "include_speaker_labels" => false,
        "split_rows" => false,
        "generate_chunks_only" => false,
        "store_file_only" => false,
        "skip_file_processing" => false,
        "parsed_text_format" => "PLAIN_TEXT",
    },
)
p result
βš™οΈ Parameters
ids: Array<AzureBlobGetFileInput>
tags: Object
chunk_size: Integer
chunk_overlap: Integer
skip_embedding_generation: Boolean
embedding_model: EmbeddingGenerators
generate_sparse_vectors: Boolean
prepend_filename_to_chunks: Boolean
max_items_per_chunk: Integer

Number of objects per chunk. For csv, tsv, xlsx, and json files only.

set_page_as_boundary: Boolean
data_source_id: Integer
request_id: String
use_ocr: Boolean
parse_pdf_tables_with_ocr: Boolean
file_sync_config: FileSyncConfigNullable
πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/integrations/azure_blob_storage/files POST

πŸ”™ Back to Table of Contents


carbon.integrations.sync_azure_blob_storage

This endpoint can be used to connect Azure Blob Storage.

For Azure Blob Storage, follow these steps:

  1. Create a new Azure Storage account and grant the following permissions:
    • List containers.
    • Read from specific containers and blobs to sync with Carbon. Ensure any future containers or blobs carry the same permissions.
  2. Generate a shared access signature (SAS) token or an access key for the storage account.

Once created, provide us with the following details to generate the connection URL:

  1. Storage Account KeyName.
  2. Storage Account Name.
πŸ› οΈ Usage
result = carbon.integrations.sync_azure_blob_storage(
  account_name: "string_example",
  account_key: "string_example",
  sync_source_items: true,
  data_source_tags: {},
)
p result
βš™οΈ Parameters
account_name: String
account_key: String
sync_source_items: Boolean
data_source_tags: Object

Tags to be associated with the data source. If the data source already has tags set, then an upsert will be performed.

πŸ”„ Return

OrganizationUserDataSourceAPI

🌐 Endpoint

/integrations/azure_blob_storage POST

πŸ”™ Back to Table of Contents


carbon.integrations.sync_confluence

Deprecated

This endpoint has been deprecated. Use /integrations/files/sync instead.

After listing pages in a user's Confluence account, the set of selected page ids and the connected account's data_source_id can be passed into this endpoint to sync them into Carbon. Additional parameters listed below can be used to associate data to the selected pages or alter the behavior of the sync.

πŸ› οΈ Usage
result = carbon.integrations.sync_confluence(
  data_source_id: 1,
  ids: [
        "string_example"
    ],
  tags: {},
  chunk_size: 1500,
  chunk_overlap: 20,
  skip_embedding_generation: false,
  embedding_model: "OPENAI",
  generate_sparse_vectors: false,
  prepend_filename_to_chunks: false,
  max_items_per_chunk: 1,
  set_page_as_boundary: false,
  request_id: "string_example",
  use_ocr: false,
  parse_pdf_tables_with_ocr: false,
  incremental_sync: false,
  file_sync_config: {
        "auto_synced_source_types" => ["ARTICLE"],
        "sync_attachments" => false,
        "detect_audio_language" => false,
        "transcription_service" => "assemblyai",
        "include_speaker_labels" => false,
        "split_rows" => false,
        "generate_chunks_only" => false,
        "store_file_only" => false,
        "skip_file_processing" => false,
        "parsed_text_format" => "PLAIN_TEXT",
    },
)
p result
βš™οΈ Parameters
data_source_id: Integer
ids: IdsProperty
tags: Object
chunk_size: Integer
chunk_overlap: Integer
skip_embedding_generation: Boolean
embedding_model: EmbeddingGeneratorsNullable
generate_sparse_vectors: Boolean
prepend_filename_to_chunks: Boolean
max_items_per_chunk: Integer

Number of objects per chunk. For csv, tsv, xlsx, and json files only.

set_page_as_boundary: Boolean
request_id: String
use_ocr: Boolean
parse_pdf_tables_with_ocr: Boolean
incremental_sync: Boolean

Only sync files if they have not already been synced or if the embedding properties have changed. This flag is currently supported by ONEDRIVE, GOOGLE_DRIVE, BOX, DROPBOX, INTERCOM, GMAIL, OUTLOOK, ZENDESK, CONFLUENCE, NOTION, SHAREPOINT, SERVICENOW. It will be ignored for other data sources.

file_sync_config: FileSyncConfigNullable
πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/integrations/confluence/sync POST

πŸ”™ Back to Table of Contents


carbon.integrations.sync_data_source_items

Sync Data Source Items

πŸ› οΈ Usage
result = carbon.integrations.sync_data_source_items(
  data_source_id: 1,
)
p result
βš™οΈ Parameters
data_source_id: Integer
πŸ”„ Return

OrganizationUserDataSourceAPI

🌐 Endpoint

/integrations/items/sync POST

πŸ”™ Back to Table of Contents


carbon.integrations.sync_files

After listing files and folders via /integrations/items/sync and integrations/items/list, use the selected items' external ids as the ids in this endpoint to sync them into Carbon. Sharepoint items take an additional parameter root_id, which identifies the drive the file or folder is in and is stored in root_external_id. That additional paramter is optional and excluding it will tell the sync to assume the item is stored in the default Documents drive.

πŸ› οΈ Usage
result = carbon.integrations.sync_files(
  data_source_id: 1,
  ids: [
        "string_example"
    ],
  tags: {},
  chunk_size: 1500,
  chunk_overlap: 20,
  skip_embedding_generation: false,
  embedding_model: "OPENAI",
  generate_sparse_vectors: false,
  prepend_filename_to_chunks: false,
  max_items_per_chunk: 1,
  set_page_as_boundary: false,
  request_id: "string_example",
  use_ocr: false,
  parse_pdf_tables_with_ocr: false,
  incremental_sync: false,
  file_sync_config: {
        "auto_synced_source_types" => ["ARTICLE"],
        "sync_attachments" => false,
        "detect_audio_language" => false,
        "transcription_service" => "assemblyai",
        "include_speaker_labels" => false,
        "split_rows" => false,
        "generate_chunks_only" => false,
        "store_file_only" => false,
        "skip_file_processing" => false,
        "parsed_text_format" => "PLAIN_TEXT",
    },
)
p result
βš™οΈ Parameters
data_source_id: Integer
ids: IdsProperty
tags: Object
chunk_size: Integer
chunk_overlap: Integer
skip_embedding_generation: Boolean
embedding_model: EmbeddingGeneratorsNullable
generate_sparse_vectors: Boolean
prepend_filename_to_chunks: Boolean
max_items_per_chunk: Integer

Number of objects per chunk. For csv, tsv, xlsx, and json files only.

set_page_as_boundary: Boolean
request_id: String
use_ocr: Boolean
parse_pdf_tables_with_ocr: Boolean
incremental_sync: Boolean

Only sync files if they have not already been synced or if the embedding properties have changed. This flag is currently supported by ONEDRIVE, GOOGLE_DRIVE, BOX, DROPBOX, INTERCOM, GMAIL, OUTLOOK, ZENDESK, CONFLUENCE, NOTION, SHAREPOINT, SERVICENOW. It will be ignored for other data sources.

file_sync_config: FileSyncConfigNullable
πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/integrations/files/sync POST

πŸ”™ Back to Table of Contents


carbon.integrations.sync_git_hub

Refer this article to obtain an access token https://docs.github.com/en/authentication/keeping-your-account-and-data-secure/managing-your-personal-access-tokens. Make sure that your access token has the permission to read content from your desired repos. Note that if your access token expires you will need to manually update it through this endpoint.

πŸ› οΈ Usage
result = carbon.integrations.sync_git_hub(
  username: "string_example",
  access_token: "string_example",
  sync_source_items: false,
  data_source_tags: {},
)
p result
βš™οΈ Parameters
username: String
access_token: String
sync_source_items: Boolean

Enabling this flag will fetch all available content from the source to be listed via list items endpoint

data_source_tags: Object

Tags to be associated with the data source. If the data source already has tags set, then an upsert will be performed.

πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/integrations/github POST

πŸ”™ Back to Table of Contents


carbon.integrations.sync_gitbook

You can sync upto 20 Gitbook spaces at a time using this endpoint. Additional parameters below can be used to associate data with the synced pages or modify the sync behavior.

πŸ› οΈ Usage
result = carbon.integrations.sync_gitbook(
  space_ids: [
        "string_example"
    ],
  data_source_id: 1,
  tags: {},
  chunk_size: 1500,
  chunk_overlap: 20,
  skip_embedding_generation: false,
  embedding_model: "OPENAI",
  generate_sparse_vectors: false,
  prepend_filename_to_chunks: false,
  request_id: "string_example",
  file_sync_config: {
        "auto_synced_source_types" => ["ARTICLE"],
        "sync_attachments" => false,
        "detect_audio_language" => false,
        "transcription_service" => "assemblyai",
        "include_speaker_labels" => false,
        "split_rows" => false,
        "generate_chunks_only" => false,
        "store_file_only" => false,
        "skip_file_processing" => false,
        "parsed_text_format" => "PLAIN_TEXT",
    },
)
p result
βš™οΈ Parameters
space_ids: Array<String>
data_source_id: Integer
tags: Object
chunk_size: Integer
chunk_overlap: Integer
skip_embedding_generation: Boolean
embedding_model: EmbeddingGenerators
generate_sparse_vectors: Boolean
prepend_filename_to_chunks: Boolean
request_id: String
file_sync_config: FileSyncConfigNullable
🌐 Endpoint

/integrations/gitbook/sync POST

πŸ”™ Back to Table of Contents


carbon.integrations.sync_gmail

Once you have successfully connected your gmail account, you can choose which emails to sync with us using the filters parameter. Filters is a JSON object with key value pairs. It also supports AND and OR operations. For now, we support a limited set of keys listed below.

label: Inbuilt Gmail labels, for example "Important" or a custom label you created.
after or before: A date in YYYY/mm/dd format (example 2023/12/31). Gets emails after/before a certain date. You can also use them in combination to get emails from a certain period.
is: Can have the following values - starred, important, snoozed, and unread
from: Email address of the sender
to: Email address of the recipient
in: Can have the following values - sent (sync emails sent by the user)
has: Can have the following values - attachment (sync emails that have attachments)

Using keys or values outside of the specified values can lead to unexpected behaviour.

An example of a basic query with filters can be

{
    "filters": {
            "key": "label",
            "value": "Test"
        }
}

Which will list all emails that have the label "Test".

You can use AND and OR operation in the following way:

{
    "filters": {
        "AND": [
            {
                "key": "after",
                "value": "2024/01/07"
            },
            {
                "OR": [
                    {
                        "key": "label",
                        "value": "Personal"
                    },
                    {
                        "key": "is",
                        "value": "starred"
                    }
                ]
            }
        ]
    }
}

This will return emails after 7th of Jan that are either starred or have the label "Personal". Note that this is the highest level of nesting we support, i.e. you can't add more AND/OR filters within the OR filter in the above example.

πŸ› οΈ Usage
result = carbon.integrations.sync_gmail(
  filters: {},
  tags: {},
  chunk_size: 1500,
  chunk_overlap: 20,
  skip_embedding_generation: false,
  embedding_model: "OPENAI",
  generate_sparse_vectors: false,
  prepend_filename_to_chunks: false,
  data_source_id: 1,
  request_id: "string_example",
  sync_attachments: false,
  file_sync_config: {
        "auto_synced_source_types" => ["ARTICLE"],
        "sync_attachments" => false,
        "detect_audio_language" => false,
        "transcription_service" => "assemblyai",
        "include_speaker_labels" => false,
        "split_rows" => false,
        "generate_chunks_only" => false,
        "store_file_only" => false,
        "skip_file_processing" => false,
        "parsed_text_format" => "PLAIN_TEXT",
    },
  incremental_sync: false,
)
p result
βš™οΈ Parameters
filters: Object
tags: Object
chunk_size: Integer
chunk_overlap: Integer
skip_embedding_generation: Boolean
embedding_model: EmbeddingGenerators
generate_sparse_vectors: Boolean
prepend_filename_to_chunks: Boolean
data_source_id: Integer
request_id: String
sync_attachments: Boolean
file_sync_config: FileSyncConfigNullable
incremental_sync: Boolean
πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/integrations/gmail/sync POST

πŸ”™ Back to Table of Contents


carbon.integrations.sync_outlook

Once you have successfully connected your Outlook account, you can choose which emails to sync with us using the filters and folder parameter. "folder" should be the folder you want to sync from Outlook. By default we get messages from your inbox folder.
Filters is a JSON object with key value pairs. It also supports AND and OR operations. For now, we support a limited set of keys listed below.

category: Custom categories that you created in Outlook.
after or before: A date in YYYY/mm/dd format (example 2023/12/31). Gets emails after/before a certain date. You can also use them in combination to get emails from a certain period.
is: Can have the following values: flagged
from: Email address of the sender

An example of a basic query with filters can be

{
    "filters": {
            "key": "category",
            "value": "Test"
        }
}

Which will list all emails that have the category "Test".

Specifying a custom folder in the same query

{
    "folder": "Folder Name",
    "filters": {
            "key": "category",
            "value": "Test"
        }
}

You can use AND and OR operation in the following way:

{
    "filters": {
        "AND": [
            {
                "key": "after",
                "value": "2024/01/07"
            },
            {
                "OR": [
                    {
                        "key": "category",
                        "value": "Personal"
                    },
                    {
                        "key": "category",
                        "value": "Test"
                    },
                ]
            }
        ]
    }
}

This will return emails after 7th of Jan that have either Personal or Test as category. Note that this is the highest level of nesting we support, i.e. you can't add more AND/OR filters within the OR filter in the above example.

πŸ› οΈ Usage
result = carbon.integrations.sync_outlook(
  filters: {},
  tags: {},
  folder: "Inbox",
  chunk_size: 1500,
  chunk_overlap: 20,
  skip_embedding_generation: false,
  embedding_model: "OPENAI",
  generate_sparse_vectors: false,
  prepend_filename_to_chunks: false,
  data_source_id: 1,
  request_id: "string_example",
  sync_attachments: false,
  file_sync_config: {
        "auto_synced_source_types" => ["ARTICLE"],
        "sync_attachments" => false,
        "detect_audio_language" => false,
        "transcription_service" => "assemblyai",
        "include_speaker_labels" => false,
        "split_rows" => false,
        "generate_chunks_only" => false,
        "store_file_only" => false,
        "skip_file_processing" => false,
        "parsed_text_format" => "PLAIN_TEXT",
    },
  incremental_sync: false,
)
p result
βš™οΈ Parameters
filters: Object
tags: Object
folder: String
chunk_size: Integer
chunk_overlap: Integer
skip_embedding_generation: Boolean
embedding_model: EmbeddingGenerators
generate_sparse_vectors: Boolean
prepend_filename_to_chunks: Boolean
data_source_id: Integer
request_id: String
sync_attachments: Boolean
file_sync_config: FileSyncConfigNullable
incremental_sync: Boolean
πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/integrations/outlook/sync POST

πŸ”™ Back to Table of Contents


carbon.integrations.sync_repos

You can retreive repos your token has access to using /integrations/github/repos and sync their content. You can also pass full name of any public repository (username/repo-name). This will store the repo content with carbon which can be accessed through /integrations/items/list endpoint. Maximum of 25 repositories are accepted per request.

πŸ› οΈ Usage
result = carbon.integrations.sync_repos(
  repos: [
        "string_example"
    ],
  data_source_id: 1,
)
p result
βš™οΈ Parameters
repos: Array<String>
data_source_id: Integer
🌐 Endpoint

/integrations/github/sync_repos POST

πŸ”™ Back to Table of Contents


carbon.integrations.sync_rss_feed

Rss Feed

πŸ› οΈ Usage
result = carbon.integrations.sync_rss_feed(
  url: "string_example",
  tags: {},
  chunk_size: 1500,
  chunk_overlap: 20,
  skip_embedding_generation: false,
  embedding_model: "OPENAI",
  generate_sparse_vectors: false,
  prepend_filename_to_chunks: false,
  request_id: "string_example",
  data_source_tags: {},
)
p result
βš™οΈ Parameters
url: String
tags: Object
chunk_size: Integer
chunk_overlap: Integer
skip_embedding_generation: Boolean
embedding_model: EmbeddingGenerators
generate_sparse_vectors: Boolean
prepend_filename_to_chunks: Boolean
request_id: String
data_source_tags: Object

Tags to be associated with the data source. If the data source already has tags set, then an upsert will be performed.

πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/integrations/rss_feed POST

πŸ”™ Back to Table of Contents


carbon.integrations.sync_s3_files

After optionally loading the items via /integrations/items/sync and integrations/items/list, use the bucket name and object key as the ID in this endpoint to sync them into Carbon. Additional parameters below can associate data with the selected items or modify the sync behavior

πŸ› οΈ Usage
result = carbon.integrations.sync_s3_files(
  ids: [
        {
        }
    ],
  tags: {},
  chunk_size: 1500,
  chunk_overlap: 20,
  skip_embedding_generation: false,
  embedding_model: "OPENAI",
  generate_sparse_vectors: false,
  prepend_filename_to_chunks: false,
  max_items_per_chunk: 1,
  set_page_as_boundary: false,
  data_source_id: 1,
  request_id: "string_example",
  use_ocr: false,
  parse_pdf_tables_with_ocr: false,
  file_sync_config: {
        "auto_synced_source_types" => ["ARTICLE"],
        "sync_attachments" => false,
        "detect_audio_language" => false,
        "transcription_service" => "assemblyai",
        "include_speaker_labels" => false,
        "split_rows" => false,
        "generate_chunks_only" => false,
        "store_file_only" => false,
        "skip_file_processing" => false,
        "parsed_text_format" => "PLAIN_TEXT",
    },
)
p result
βš™οΈ Parameters
ids: Array<S3GetFileInput>

Each input should be one of the following: A bucket name, a bucket name and a prefix, or a bucket name and an object key. A prefix is the common path for all objects you want to sync. Paths should end with a forward slash.

tags: Object
chunk_size: Integer
chunk_overlap: Integer
skip_embedding_generation: Boolean
embedding_model: EmbeddingGenerators
generate_sparse_vectors: Boolean
prepend_filename_to_chunks: Boolean
max_items_per_chunk: Integer

Number of objects per chunk. For csv, tsv, xlsx, and json files only.

set_page_as_boundary: Boolean
data_source_id: Integer
request_id: String
use_ocr: Boolean
parse_pdf_tables_with_ocr: Boolean
file_sync_config: FileSyncConfigNullable
πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/integrations/s3/files POST

πŸ”™ Back to Table of Contents


carbon.integrations.sync_slack

You can list all conversations using the endpoint /integrations/slack/conversations. The ID of conversation will be used as an input for this endpoint with timestamps as optional filters.

πŸ› οΈ Usage
result = carbon.integrations.sync_slack(
  filters: {
        "conversation_id" => "conversation_id_example",
    },
  tags: {},
  chunk_size: 1500,
  chunk_overlap: 20,
  skip_embedding_generation: false,
  embedding_model: "OPENAI",
  generate_sparse_vectors: false,
  prepend_filename_to_chunks: false,
  data_source_id: 1,
  request_id: "string_example",
)
p result
βš™οΈ Parameters
filters: SlackFilters
tags: Object
chunk_size: Integer
chunk_overlap: Integer
skip_embedding_generation: Boolean
embedding_model: EmbeddingGenerators
generate_sparse_vectors: Boolean
prepend_filename_to_chunks: Boolean
data_source_id: Integer
request_id: String
🌐 Endpoint

/integrations/slack/sync POST

πŸ”™ Back to Table of Contents


carbon.organizations.get

Get Organization

πŸ› οΈ Usage
result = carbon.organizations.get
p result
πŸ”„ Return

OrganizationResponse

🌐 Endpoint

/organization GET

πŸ”™ Back to Table of Contents


carbon.organizations.update

Update Organization

πŸ› οΈ Usage
result = carbon.organizations.update(
  global_user_config: {
    },
  data_source_configs: {
        "key": {
            "allowed_file_formats" => [],
        },
    },
)
p result
βš™οΈ Parameters
global_user_config: UserConfigurationNullable
data_source_configs: Hash<string, DataSourceConfiguration>

Used to set organization level defaults for configuration related to data sources.

πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/organization/update POST

πŸ”™ Back to Table of Contents


carbon.organizations.update_stats

Use this endpoint to reaggregate the statistics for an organization, for example aggregate_file_size. The reaggregation process is asyncronous so a webhook will be sent with the event type being FILE_STATISTICS_AGGREGATED to notify when the process is complee. After this aggregation is complete, the updated statistics can be retrieved using the /organization endpoint. The response of /organization willalso contain a timestamp of the last time the statistics were reaggregated.

πŸ› οΈ Usage
result = carbon.organizations.update_stats
p result
πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/organization/statistics POST

πŸ”™ Back to Table of Contents


carbon.users.delete

Delete Users

πŸ› οΈ Usage
result = carbon.users.delete(
  customer_ids: [
        "string_example"
    ],
)
p result
βš™οΈ Parameters
customer_ids: Array<String>
πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/delete_users POST

πŸ”™ Back to Table of Contents


carbon.users.get

User Endpoint

πŸ› οΈ Usage
result = carbon.users.get(
  customer_id: "string_example",
)
p result
βš™οΈ Parameters
customer_id: String
πŸ”„ Return

UserResponse

🌐 Endpoint

/user POST

πŸ”™ Back to Table of Contents


carbon.users.list

List users within an organization

πŸ› οΈ Usage
result = carbon.users.list(
  pagination: {
        "limit" => 10,
        "offset" => 0,
        "starting_id" => 0,
    },
  filters: {
    },
  order_by: "created_at",
  order_dir: "asc",
  include_count: false,
)
p result
βš™οΈ Parameters
pagination: Pagination
filters: ListUsersFilters
order_by: ListUsersOrderByTypes
order_dir: OrderDirV2
include_count: Boolean
πŸ”„ Return

UserListResponse

🌐 Endpoint

/list_users POST

πŸ”™ Back to Table of Contents


carbon.users.toggle_user_features

Deprecated

Toggle User Features

πŸ› οΈ Usage
result = carbon.users.toggle_user_features(
  configuration_key_name: "sparse_vectors",
  value: {},
)
p result
βš™οΈ Parameters
configuration_key_name: ConfigurationKeys
value: Object
πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/modify_user_configuration POST

πŸ”™ Back to Table of Contents


carbon.users.update_users

Update Users

πŸ› οΈ Usage
result = carbon.users.update_users(
  customer_ids: [
        "string_example"
    ],
  auto_sync_enabled_sources: [
        "string_example"
    ],
  max_files: -1,
  max_files_per_upload: -1,
  max_characters: -1,
  max_characters_per_file: -1,
  max_characters_per_upload: -1,
  auto_sync_interval: -1,
)
p result
βš™οΈ Parameters
customer_ids: Array<String>

List of organization supplied user IDs

auto_sync_enabled_sources: AutoSyncEnabledSourcesProperty
max_files: Integer

Custom file upload limit for the user over all user's files across all uploads. If set, then the user will not be allowed to upload more files than this limit. If not set, or if set to -1, then the user will have no limit.

max_files_per_upload: Integer

Custom file upload limit for the user across a single upload. If set, then the user will not be allowed to upload more files than this limit in a single upload. If not set, or if set to -1, then the user will have no limit.

max_characters: Integer

Custom character upload limit for the user over all user's files across all uploads. If set, then the user will not be allowed to upload more characters than this limit. If not set, or if set to -1, then the user will have no limit.

max_characters_per_file: Integer

A single file upload from the user can not exceed this character limit. If set, then the file will not be synced if it exceeds this limit. If not set, or if set to -1, then the user will have no limit.

max_characters_per_upload: Integer

Custom character upload limit for the user across a single upload. If set, then the user won't be able to sync more than this many characters in one upload. If not set, or if set to -1, then the user will have no limit.

auto_sync_interval: Integer

The interval in hours at which the user's data sources should be synced. If not set or set to -1, the user will be synced at the organization level interval or default interval if that is also not set. Must be one of [3, 6, 12, 24]

πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/update_users POST

πŸ”™ Back to Table of Contents


carbon.users.who_am_i

Me Endpoint

πŸ› οΈ Usage
result = carbon.users.who_am_i
p result
πŸ”„ Return

UserResponse

🌐 Endpoint

/whoami GET

πŸ”™ Back to Table of Contents


carbon.utilities.fetch_urls

Deprecated

Extracts all URLs from a webpage.

Args: url (str): URL of the webpage

Returns: FetchURLsResponse: A response object with a list of URLs extracted from the webpage and the webpage content.

πŸ› οΈ Usage
result = carbon.utilities.fetch_urls(
  url: "url_example",
)
p result
βš™οΈ Parameters
url: String
πŸ”„ Return

FetchURLsResponse

🌐 Endpoint

/fetch_urls GET

πŸ”™ Back to Table of Contents


carbon.utilities.fetch_webpage

Fetch Urls V2

πŸ› οΈ Usage
result = carbon.utilities.fetch_webpage(
  url: "string_example",
)
p result
βš™οΈ Parameters
url: String
🌐 Endpoint

/fetch_webpage POST

πŸ”™ Back to Table of Contents


carbon.utilities.fetch_youtube_transcripts

Fetches english transcripts from YouTube videos.

Args: id (str): The ID of the YouTube video. raw (bool): Whether to return the raw transcript or not. Defaults to False.

Returns: dict: A dictionary with the transcript of the YouTube video.

πŸ› οΈ Usage
result = carbon.utilities.fetch_youtube_transcripts(
  id: "id_example",
  raw: false,
)
p result
βš™οΈ Parameters
id: String
raw: Boolean
πŸ”„ Return

YoutubeTranscriptResponse

🌐 Endpoint

/fetch_youtube_transcript GET

πŸ”™ Back to Table of Contents


carbon.utilities.process_sitemap

Retrieves all URLs from a sitemap, which can subsequently be utilized with our web_scrape endpoint.

πŸ› οΈ Usage
result = carbon.utilities.process_sitemap(
  url: "url_example",
)
p result
βš™οΈ Parameters
url: String
🌐 Endpoint

/process_sitemap GET

πŸ”™ Back to Table of Contents


carbon.utilities.scrape_sitemap

Extracts all URLs from a sitemap and performs a web scrape on each of them.

Args: sitemap_url (str): URL of the sitemap

Returns: dict: A response object with the status of the scraping job message.-->

πŸ› οΈ Usage
result = carbon.utilities.scrape_sitemap(
  url: "string_example",
  tags: {
        "key": "string_example",
    },
  max_pages_to_scrape: 1,
  chunk_size: 1500,
  chunk_overlap: 20,
  skip_embedding_generation: false,
  enable_auto_sync: false,
  generate_sparse_vectors: false,
  prepend_filename_to_chunks: false,
  html_tags_to_skip: [],
  css_classes_to_skip: [],
  css_selectors_to_skip: [],
  embedding_model: "OPENAI",
  url_paths_to_include: [],
  url_paths_to_exclude: [],
  urls_to_scrape: [],
  download_css_and_media: false,
  generate_chunks_only: false,
  store_file_only: false,
  use_premium_proxies: false,
)
p result
βš™οΈ Parameters
url: String
tags: Hash<string, Tags1>
max_pages_to_scrape: Integer
chunk_size: Integer
chunk_overlap: Integer
skip_embedding_generation: Boolean
enable_auto_sync: Boolean
generate_sparse_vectors: Boolean
prepend_filename_to_chunks: Boolean
html_tags_to_skip: Array<String>
css_classes_to_skip: Array<String>
css_selectors_to_skip: Array<String>
embedding_model: EmbeddingGenerators
url_paths_to_include: Array<String>

URL subpaths or directories that you want to include. For example if you want to only include URLs that start with /questions in stackoverflow.com, you will add /questions/ in this input

url_paths_to_exclude: Array<String>

URL subpaths or directories that you want to exclude. For example if you want to exclude URLs that start with /questions in stackoverflow.com, you will add /questions/ in this input

urls_to_scrape: Array<String>

You can submit a subset of URLs from the sitemap that should be scraped. To get the list of URLs, you can check out /process_sitemap endpoint. If left empty, all URLs from the sitemap will be scraped.

download_css_and_media: Boolean

Whether the scraper should download css and media from the page (images, fonts, etc). Scrapes might take longer to finish with this flag enabled, but the success rate is improved.

generate_chunks_only: Boolean

If this flag is enabled, the file will be chunked and stored with Carbon, but no embeddings will be generated. This overrides the skip_embedding_generation flag.

store_file_only: Boolean

If this flag is enabled, the file will be stored with Carbon, but no processing will be done.

use_premium_proxies: Boolean

If the default proxies are blocked and not returning results, this flag can be enabled to use alternate proxies (residential and office). Scrapes might take longer to finish with this flag enabled.

🌐 Endpoint

/scrape_sitemap POST

πŸ”™ Back to Table of Contents


carbon.utilities.scrape_web

Conduct a web scrape on a given webpage URL. Our web scraper is fully compatible with JavaScript and supports recursion depth, enabling you to efficiently extract all content from the target website.

πŸ› οΈ Usage
result = carbon.utilities.scrape_web(
  body: [
        {
            "url" => "url_example",
            "recursion_depth" => 3,
            "max_pages_to_scrape" => 100,
            "chunk_size" => 1500,
            "chunk_overlap" => 20,
            "skip_embedding_generation" => false,
            "enable_auto_sync" => false,
            "generate_sparse_vectors" => false,
            "prepend_filename_to_chunks" => false,
            "html_tags_to_skip" => [],
            "css_classes_to_skip" => [],
            "css_selectors_to_skip" => [],
            "embedding_model" => "OPENAI",
            "url_paths_to_include" => [],
            "download_css_and_media" => false,
            "generate_chunks_only" => false,
            "store_file_only" => false,
            "use_premium_proxies" => false,
        }
    ],
)
p result
βš™οΈ body

Array<WebscrapeRequest>

🌐 Endpoint

/web_scrape POST

πŸ”™ Back to Table of Contents


carbon.utilities.search_urls

Perform a web search and obtain a list of relevant URLs.

As an illustration, when you perform a search for β€œcontent related to MRNA,” you will receive a list of links such as the following:

- https://tomrenz.substack.com/p/mrna-and-why-it-matters

- https://www.statnews.com/2020/11/10/the-story-of-mrna-how-a-once-dismissed-idea-became-a-leading-technology-in-the-covid-vaccine-race/

- https://www.statnews.com/2022/11/16/covid-19-vaccines-were-a-success-but-mrna-still-has-a-delivery-problem/

- https://joomi.substack.com/p/were-still-being-misled-about-how

Subsequently, you can submit these links to the web_scrape endpoint in order to retrieve the content of the respective web pages.

Args: query (str): Query to search for

Returns: FetchURLsResponse: A response object with a list of URLs for a given search query.

πŸ› οΈ Usage
result = carbon.utilities.search_urls(
  query: "query_example",
)
p result
βš™οΈ Parameters
query: String
πŸ”„ Return

FetchURLsResponse

🌐 Endpoint

/search_urls GET

πŸ”™ Back to Table of Contents


carbon.utilities.user_webpages

User Web Pages

πŸ› οΈ Usage
result = carbon.utilities.user_webpages(
  filters: {
    },
  pagination: {
        "limit" => 10,
        "offset" => 0,
        "starting_id" => 0,
    },
  order_by: "created_at",
  order_dir: "asc",
)
p result
βš™οΈ Parameters
filters: UserWebPagesFilters
pagination: Pagination
order_by: UserWebPageOrderByTypes
order_dir: OrderDirV2
🌐 Endpoint

/user_webpages POST

πŸ”™ Back to Table of Contents


carbon.webhooks.add_url

Add Webhook Url

πŸ› οΈ Usage
result = carbon.webhooks.add_url(
  url: "string_example",
)
p result
βš™οΈ Parameters
url: String
πŸ”„ Return

Webhook

🌐 Endpoint

/add_webhook POST

πŸ”™ Back to Table of Contents


carbon.webhooks.delete_url

Delete Webhook Url

πŸ› οΈ Usage
result = carbon.webhooks.delete_url(
  webhook_id: 1,
)
p result
βš™οΈ Parameters
webhook_id: Integer
πŸ”„ Return

GenericSuccessResponse

🌐 Endpoint

/delete_webhook/{webhook_id} DELETE

πŸ”™ Back to Table of Contents


carbon.webhooks.urls

Webhook Urls

πŸ› οΈ Usage
result = carbon.webhooks.urls(
  pagination: {
        "limit" => 10,
        "offset" => 0,
        "starting_id" => 0,
    },
  order_by: "created_at",
  order_dir: "desc",
  filters: {
        "ids" => [],
    },
)
p result
βš™οΈ Parameters
pagination: Pagination
order_by: WebhookOrderByColumns
order_dir: OrderDir
filters: WebhookFilters
πŸ”„ Return

WebhookQueryResponse

🌐 Endpoint

/webhooks POST

πŸ”™ Back to Table of Contents


carbon.white_label.create

Create White Labels

πŸ› οΈ Usage
result = carbon.white_label.create(
  body: [
        {
            "data_source_type" => "GOOGLE_DRIVE",
            "credentials" => {
                "client_id" => "client_id_example",
                "redirect_uri" => "redirect_uri_example",
            },
        }
    ],
)
p result
βš™οΈ body

Array<WhiteLabelCreateRequestInner>

🌐 Endpoint

/white_label/create POST

πŸ”™ Back to Table of Contents


carbon.white_label.delete

Delete White Labels

πŸ› οΈ Usage
result = carbon.white_label.delete(
  ids: [
        1
    ],
)
p result
βš™οΈ Parameters
ids: Array<Integer>
🌐 Endpoint

/white_label/delete POST

πŸ”™ Back to Table of Contents


carbon.white_label.list

List White Labels

πŸ› οΈ Usage
result = carbon.white_label.list(
  pagination: {
        "limit" => 10,
        "offset" => 0,
        "starting_id" => 0,
    },
  order_by: "created_at",
  order_dir: "desc",
  filters: {
        "ids" => [],
        "data_source_type" => [],
    },
)
p result
βš™οΈ Parameters
pagination: Pagination
order_by: WhiteLabelOrderByColumns
order_dir: OrderDir
filters: WhiteLabelFilters
🌐 Endpoint

/white_label/list POST

πŸ”™ Back to Table of Contents


carbon.white_label.update

Update White Label

πŸ› οΈ Usage
result = carbon.white_label.update(
  data_source_type: "INTERCOM",
  credentials: {
        "client_id" => "client_id_example",
        "redirect_uri" => "redirect_uri_example",
    },
  body: {
        "data_source_type" => "GOOGLE_DRIVE",
        "credentials" => {
            "client_id" => "client_id_example",
            "redirect_uri" => "redirect_uri_example",
        },
    },
)
p result
βš™οΈ Parameters
data_source_type: WhiteLabelInputDataSourceType
credentials: Credentials
🌐 Endpoint

/white_label/update POST

πŸ”™ Back to Table of Contents


Author

This TypeScript package is automatically generated by Konfig

FAQs

Package last updated on 02 Dec 2024

Did you know?

Socket

Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.

Install

Related posts

SocketSocket SOC 2 Logo

Product

  • Package Alerts
  • Integrations
  • Docs
  • Pricing
  • FAQ
  • Roadmap
  • Changelog

Packages

npm

Stay in touch

Get open source security insights delivered straight into your inbox.


  • Terms
  • Privacy
  • Security

Made with ⚑️ by Socket Inc