Huge News!Announcing our $40M Series B led by Abstract Ventures.Learn More
Socket
Sign inDemoInstall
Socket

ai.ancf.lmos-router:lmos-router-vector-spring-boot-starter

Package Overview
Dependencies
Maintainers
3
Alerts
File Explorer

Advanced tools

Socket logo

Install Socket

Detect and block malicious and high-risk dependencies

Install

ai.ancf.lmos-router:lmos-router-vector-spring-boot-starter

Efficient Agent Routing with SOTA Language and Embedding Models.

  • 0.2.0
  • Source
  • Maven
  • Socket score

Version published
Maintainers
3
Source

Intelligent Agent Routing System

Quickstart Guide

This quickstart guide will help you set up and use the agent routing system. Follow the steps below to get started.

Step 1: Create Agent Routing Specifications

First, download lmos-router dependency from maven central

implementation("ai.ancf.lmos:lmos-router-llm:x.y.z")

then create the agent routing specifications using SimpleAgentRoutingSpecProvider and AgentRoutingSpecBuilder.

val agentRoutingSpecsProvider = SimpleAgentRoutingSpecProvider()
    .add(
        AgentRoutingSpecBuilder()
            .name("offer-agent")
            .description("This agent is responsible for offer management")
            .version("1.0.0")
            .address(Address(uri = "/agents/offer-agent"))
            .build()
    )
    .add(
        AgentRoutingSpecBuilder()
            .name("service-agent")
            .description("This agent is responsible for service management")
            .version("1.0.0")
            .address(Address(uri = "/agents/service-agent"))
            .build()
    )

Step 2: Initialize the Agent Routing Spec Resolver

Next, initialize the LLMAgentRoutingSpecsResolver with the agentRoutingSpecsProvider.

val agentRoutingSpecResolver = LLMAgentRoutingSpecsResolver(
    agentRoutingSpecsProvider,
    modelClient = DefaultModelClient(
        DefaultModelClientProperties(openAiApiKey = "your-openai-api-key") //Defaults to System.getenv("OPENAI_API_KEY")
    )
)

Step 3: Pass the Context and User Message

Set up the context and input messages that will be used to resolve the appropriate agent.

val context = Context(listOf(AssistantMessage("Hello")))
val input = UserMessage("Can you help me find a new phone?")

Step 4: Resolve the Agent

Finally, use the agentRoutingSpecResolver to resolve the appropriate agent based on the context and input messages.

val result = agentRoutingSpecResolver.resolve(context, input)

The result should return offer-agent, indicating that the "offer-agent" is responsible for handling the user's request. Now you can use the address uri to route the user to the appropriate agent.

For spring cloud gateway, refer to the Demo.

Overview

This project routes user queries to the most suitable agent based on their capabilities using Language Model (LLM), Vector-based approaches, and a new Hybrid approach.

Table of Contents

  1. Introduction
  2. Routing Methods
  3. Modules
  4. Benchmarks
  5. Setup and Installation
  6. Demo
  7. Contributing
  8. License

Introduction

The Intelligent Agent Routing System directs user queries to the best-suited agent based on their capabilities using three methods:

  1. LLM-based approach: Uses a language model to understand and match queries with agent capabilities.
  2. Vector-based approach: Uses vector embeddings to find the most similar agent based on the query.
  3. Hybrid approach: Extracts abstract requirements from the query using an LLM and then searches for an agent using semantic similarity.

Routing Methods

LLM-Based Approach

Uses advanced language models like OpenAI's GPT-3.5 to understand the context and semantics of user queries.

Pros:

  • Understands complex queries and context.
  • Flexible and adaptable to various scenarios.
  • Utilizes state-of-the-art NLP techniques.

Cons:

  • Expensive due to commercial language model costs.
  • Higher response times.
  • Dependent on external APIs with potential rate limits.

Vector-Based Approach

Uses vector embeddings to represent queries and agent capabilities, comparing them using cosine similarity.

Pros:

  • Fast and efficient for large-scale data.
  • Scalable to handle more agents and queries.
  • Independent of external APIs.

Cons:

  • Limited in understanding complex queries.
  • Requires initial setup and regular updates.
  • Needs maintenance for embedding updates.

Hybrid Approach

Extracts abstract requirements from the query using an LLM and then searches for an agent using semantic similarity.

Pros:

  • Balances the strengths of both LLM and Vector-based approaches.
  • Better understanding of complex queries than vector-based alone.
  • More efficient than LLM-based alone.

Cons:

  • Still dependent on external APIs for LLM.
  • Requires integration of both LLM and vector-based systems.

Comparison Table

FeatureLLM-Based ApproachVector-Based ApproachHybrid Approach
Contextual UnderstandingHighModerateHigh
FlexibilityHighModerateHigh
EfficiencyModerateHighHigh
ScalabilityModerateHighHigh
CostHighLowHigh
LatencyHigherLowerHigh
DependencyHighLowHigh
Setup ComplexityLowHighHigh
MaintenanceLowHighHigh

Modules

Core

Contains foundational classes and interfaces:

  • ChatMessage: Represents different types of chat messages.
  • Context: Represents the conversation context.
  • AgentRoutingSpec: Represents agent routing specifications.
  • AgentRoutingSpecsProvider: Interface for providing agent routing specifications.
  • AgentRoutingSpecsResolver: Interface for resolving agent routing specifications.
  • Result: Utility class for handling success and failure cases.

LLM

Handles agent routing specifications using a language model:

  • DefaultModelClient: Client for calling the OpenAI model.
  • LLMAgentRoutingSpecsResolver: Resolves agent routing specifications using a language model.
  • ModelPromptProvider: Provides prompts for the language model.

Vector

Handles agent routing specifications using vector embeddings:

  • DefaultEmbeddingClient: Client for embedding text using a local service.
  • OpenAIEmbeddingClient: Client for embedding text using the OpenAI API.
  • VectorAgentRoutingSpecsResolver: Resolves agent routing specifications using vector similarity search.
  • VectorSearchClient: Interface for searching similar vectors.
  • VectorSeedClient: Interface for seeding vectors.

Hybrid

Combines LLM and vector-based approaches:

  • HybridAgentRoutingSpecsResolver: Resolves agent routing specifications using a hybrid approach.

LLM Spring boot starter

Spring Boot starter for the LLM-based agent routing system:

  • LLMAgentRoutingSpecsResolverAutoConfiguration: Auto-configuration for the LLM-based agent routing system.
  • LLMAgentRoutingSpecsResolverProperties: Configuration properties for the LLM-based agent routing system.
  • LLMAgentRoutingSpecsResolverService: Service for resolving agent routing specifications.

Vector Spring boot starter

Spring Boot starter for the Vector-based agent routing system:

  • VectorAgentRoutingSpecsResolverAutoConfiguration: Auto-configuration for the Vector-based agent routing system.
  • VectorAgentRoutingSpecsResolverProperties: Configuration properties for the Vector-based agent routing system.
  • VectorAgentRoutingSpecsResolverService: Service for resolving agent routing specifications.
  • VectorSeedService: Service for seeding vectors.
  • VectorSearchService: Service for searching similar vectors.

Hybrid Spring boot starter

Spring Boot starter for the Hybrid-based agent routing system:

  • HybridAgentRoutingSpecsResolverAutoConfiguration: Auto-configuration for the Hybrid-based agent routing system.
  • HybridAgentRoutingSpecsResolverProperties: Configuration properties for the Hybrid-based agent routing system.
  • HybridAgentRoutingSpecsResolverService: Service for resolving agent routing specifications.

Demo

Sample Spring Boot application demonstrating the system:

  • AgentsApplication: Main application class.
  • AgentsController: REST controller for handling agent responses.
  • SuperRouteGatewayApplication: Spring Cloud Gateway application for routing requests.

Benchmarks

Evaluates the performance of the LLM-based, Vector-based, and Hybrid resolvers:

  • LLM-based Resolver: Processes 2000 samples.
  • Vector-based Resolver: Processes 5000 samples.
  • Hybrid Resolver: To be added.

Refer to the Benchmarks for detailed instructions.

Confusion Matrix and Accuracy

The benchmarks include confusion matrices and accuracy metrics for all methods.

LLM-Based Resolver

LLM Confusion Matrix LLM Accuracy

Vector-Based Resolver

Vector Confusion Matrix Vector Accuracy

Setup and Installation

You can download the dependencies from maven central by adding the following dependencies to your project:

LLM-Based Approach Spring Boot Starter

implementation("ai.ancf.lmos:lmos-router-llm-spring-boot-starter:x.y.z")

Or using Maven:

<dependency>
    <groupId>ai.ancf.lmos</groupId>
    <artifactId>lmos-router-llm-spring-boot-starter</artifactId>
    <version>x.y.z</version>
</dependency>

Vector-Based Approach Spring Boot Starter

implementation("ai.ancf.lmos:lmos-router-vector-spring-boot-starter:x.y.z")

Or using Maven:

<dependency>
    <groupId>ai.ancf.lmos</groupId>
    <artifactId>lmos-router-vector-spring-boot-starter</artifactId>
    <version>x.y.z</version>
</dependency>

Hybrid Approach Spring Boot Starter

implementation("ai.ancf.lmos:lmos-router-hybrid-spring-boot-starter:x.y.z")

Or using Maven:

<dependency>
    <groupId>ai.ancf.lmos</groupId>
    <artifactId>lmos-router-hybrid-spring-boot-starter</artifactId>
    <version>x.y.z</version>
</dependency>

No framework dependencies

If you are not using Spring Boot, you can add the following dependencies:

LLM-Based Approach
implementation("ai.ancf.lmos:lmos-router-llm:x.y.z")
Vector-Based Approach
implementation("ai.ancf.lmos:lmos-router-vector:x.y.z")
Hybrid Approach
implementation("ai.ancf.lmos:lmos-router-hybrid:x.y.z")

or you can build the project from source:

  1. Clone the repository:
git clone https://github.com/lmos-ai/lmos-router.git
cd lmos-router
  1. Set environment variables: (If running Flow tests, they can be enabled by setting gradle project property runFlowTests=true)

    • OPENAI_API_KEY: Your OpenAI API key.
    • VECTOR_SEED_JSON_FILE_PATH: Path to the JSON file containing seed vectors.
  2. Build the project:

./gradlew build

Demo

To run the demo:

Refer to the Demo for detailed instructions.

Contributing

Contributions are welcome! Please read the contributing guidelines for more information.

Code of Conduct

This project has adopted the Contributor Covenant in version 2.1 as our code of conduct. Please see the details in our CodeOfConduct.md. All contributors must abide by the code of conduct.

By participating in this project, you agree to abide by its Code of Conduct at all times.

Licensing

Copyright (c) 2024 Deutsche Telekom AG.

Sourcecode licensed under the Apache License, Version 2.0 (the "License"); you may not use this project except in compliance with the License.

This project follows the REUSE standard for software licensing.
Each file contains copyright and license information, and license texts can be found in the ./LICENSES folder. For more information visit https://reuse.software/.

Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the LICENSE for the specific language governing permissions and limitations under the License.

FAQs

Package last updated on 24 Sep 2024

Did you know?

Socket

Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.

Install

Related posts

SocketSocket SOC 2 Logo

Product

  • Package Alerts
  • Integrations
  • Docs
  • Pricing
  • FAQ
  • Roadmap
  • Changelog

Packages

npm

Stay in touch

Get open source security insights delivered straight into your inbox.


  • Terms
  • Privacy
  • Security

Made with ⚡️ by Socket Inc