Huge News!Announcing our $40M Series B led by Abstract Ventures.Learn More
Socket
Sign inDemoInstall
Socket

autonomize-autorag

Package Overview
Dependencies
Maintainers
1
Alerts
File Explorer

Advanced tools

Socket logo

Install Socket

Detect and block malicious and high-risk dependencies

Install

autonomize-autorag

AutoRAG is a flexible and scalable solution for building Retrieval-Augmented Generation (RAG) systems.

  • 0.1.28
  • PyPI
  • Socket score

Maintainers
1

AutoRAG

Powering seamless retrieval and generation workflows for our internal AI systems

Python Version PyPI Version Code Formatter Code Linter Code Checker Code Coverage

Overview

AutoRAG is a flexible and scalable solution for building Retrieval-Augmented Generation (RAG) systems.

This SDK provides out-of-the-box functionality for creating and managing retrieval-augmented generation workflows, offering a modular, highly-configurable interface. It supports multiple vector stores and leverages http clients like httpx for handling requests, ensuring seamless integration.

Features

  • Modular architecture: The SDK allows you to swap, extend, or customize components like retrieval models, vector stores, and response generation strategies.
  • High scalability: Built to handle large-scale data retrieval and generation, enabling robust, production-ready applications.
  • Celery for dependency injection: Efficient background tasks with support for distributed task execution.
  • Multi-flow support: Easily integrate various vector databases (ex: Qdrant, Azure AI Search) with various language models providers (ex: OpenAI, vLLM, Ollama) using standardized public methods for seamless development.

Installation

  1. Create a virtual environment, we recommend Miniconda for environment management:
    conda create -n autorag python=3.12
    conda activate autorag
    
  2. Install the package:
    pip install autonomize-autorag
    

To install with optional dependencies like Qdrant, Huggingface, OpenAI, Modelhub, etc., refer to the Installation Guide.

Usage

The full set of examples can be found in examples directory.

Sync Usage

import os
from autorag.language_models import OpenAILanguageModel

llm = OpenAILanguageModel(
    api_key=os.environ.get("OPENAI_API_KEY"),
)

generation = llm.generate(
    message=[{"role": "user", "content": "What is attention in ML?"}],
    model="gpt-4o"
)

Async Usage

Simply use sync methods with a prefix and use await for each call. Example: client.generate(...) becomes await client.agenerate(...) and everything else remains the same.

import os
from autorag.language_models import OpenAILanguageModel

llm = OpenAILanguageModel(
    api_key=os.environ.get("OPENAI_API_KEY"),
)

generation = await llm.agenerate(
    message=[{"role": "user", "content": "What is attention in ML?"}],
    model="gpt-4o"
)

Contribution

To contribute in our AutoRAG SDK, please refer to our Contribution Guidelines.

License

Copyright (C) Autonomize AI - All Rights Reserved

The contents of this repository cannot be copied and/or distributed without the explicit permission from Autonomize.ai

FAQs


Did you know?

Socket

Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.

Install

Related posts

SocketSocket SOC 2 Logo

Product

  • Package Alerts
  • Integrations
  • Docs
  • Pricing
  • FAQ
  • Roadmap
  • Changelog

Packages

npm

Stay in touch

Get open source security insights delivered straight into your inbox.


  • Terms
  • Privacy
  • Security

Made with ⚡️ by Socket Inc