New Research: Supply Chain Attack on Axios Pulls Malicious Dependency from npm.Details →
Socket
Book a DemoSign in
Socket

lance-context

Package Overview
Dependencies
Maintainers
1
Versions
76
Alerts
File Explorer

Advanced tools

Socket logo

Install Socket

Detect and block malicious and high-risk dependencies

Install

lance-context

MCP plugin for semantic code search using LanceDB - gives AI coding agents deep context from your entire codebase

alpha
latest
Source
npmnpm
Version
1.37.1
Version published
Maintainers
1
Created
Source

lance-context logo

CI npm version License: MIT Node.js version

lance-context

An MCP plugin that adds semantic code search to Claude Code and other AI coding agents, giving them deep context from your entire codebase.

Features

  • Semantic Code Search: Natural language queries locate relevant code across your entire codebase
  • Token Savings: Dramatically reduces context usage by returning only relevant code chunks
  • Multiple Embedding Backends: Google Gemini (free) or Ollama (local)
  • LanceDB Vector Storage: Fast, efficient vector search with hybrid BM25 + dense matching
  • MCP Compatible: Works with Claude Code, Cursor, and other MCP-compatible tools
  • Web Dashboard: Real-time monitoring of index status, token savings, and usage statistics
  • Beads Integration: Shows issue tracker data if your project uses beads

How lance-context Saves Tokens

AI coding agents typically need to read entire files to understand your codebase, which consumes significant context tokens. lance-context dramatically reduces token usage by:

Without lance-contextWith lance-contextSavings
Read 5-10 files to find auth code (~5000 lines)search_code returns 3 chunks (~150 lines)~97%
Read entire file to understand structureget_symbols_overview returns compact list~80-90%
Explore many files to understand codebasesummarize_codebase + list_concepts~95%
Read and compare files for duplicatessearch_similar returns targeted results~90%

Token Savings Dashboard

The web dashboard displays real-time token savings statistics:

  • Estimated Tokens Saved: Total tokens avoided by using semantic search
  • Efficiency: Percentage of potential tokens saved
  • Files Not Read: Count of files skipped due to targeted search
  • Operations Tracked: Number of search operations contributing to savings

How It Works

  • Chunking: Your codebase is split into semantic chunks (functions, classes, etc.)
  • Embedding: Each chunk is converted to a vector embedding
  • Search: Queries find only the most relevant chunks, not entire files
  • Return: Only the matching chunks are sent to the AI, saving context tokens

Installation

Add lance-context to Claude Code:

claude mcp add --scope user --transport stdio lance-context -- npx -y lance-context

Restart Claude Code to start using semantic search.

Global Install (Alternative)

For faster startup (no npm check on each run):

npm install -g lance-context

This automatically registers lance-context with Claude Code. Update manually with npm update -g lance-context.

Manual Registration

If automatic registration didn't work, manually add to Claude Code:

claude mcp add --scope user --transport stdio lance-context -- npx -y lance-context@latest

Verify Installation

In Claude Code, run /mcp to see lance-context in the list of MCP servers.

Project-Level Installation

For project-specific MCP configuration, add a .mcp.json to your project root:

{
  "mcpServers": {
    "lance-context": {
      "command": "npx",
      "args": ["-y", "lance-context@latest"]
    }
  }
}

Project Configuration

Create a .lance-context.json file in your project root to customize indexing behavior. All options are optional - lance-context works out of the box with sensible defaults.

Minimal Configuration

For most projects, you only need to specify what to include:

{
  "patterns": ["**/*.ts", "**/*.js"],
  "instructions": "This is a TypeScript monorepo. Use semantic search to find relevant utilities."
}

Full Configuration Example

{
  "patterns": ["**/*.ts", "**/*.tsx", "**/*.js", "**/*.jsx"],
  "excludePatterns": ["**/node_modules/**", "**/dist/**", "**/*.test.ts"],
  "embedding": {
    "backend": "gemini"
  },
  "chunking": {
    "maxLines": 100,
    "overlap": 20
  },
  "search": {
    "semanticWeight": 0.7,
    "keywordWeight": 0.3
  },
  "dashboard": {
    "enabled": true,
    "port": 24300,
    "openBrowser": true
  },
  "instructions": "Project-specific instructions for AI agents working with this codebase."
}

Configuration Options Reference

OptionDescriptionDefault
patternsGlob patterns for files to index["**/*.ts", "**/*.tsx", "**/*.js", "**/*.jsx", "**/*.py", "**/*.go", "**/*.rs", "**/*.java", "**/*.rb", "**/*.php", "**/*.c", "**/*.cpp", "**/*.h", "**/*.hpp", "**/*.cs", "**/*.swift", "**/*.kt"]
excludePatternsGlob patterns for files to exclude["**/node_modules/**", "**/dist/**", "**/.git/**", "**/build/**", "**/target/**", "**/__pycache__/**", "**/venv/**", "**/.venv/**", "**/vendor/**", "**/*.min.js", "**/*.min.css"]
embedding.backendEmbedding provider: "gemini" or "ollama"Auto-detect based on available API keys
embedding.modelOverride the default embedding modelBackend default
embedding.ollamaConcurrencyMax concurrent Ollama requests (1-200)100
indexing.batchSizeTexts per embedding batch request (1-1000)200
chunking.maxLinesMaximum lines per chunk100
chunking.overlapOverlapping lines between chunks for context continuity20
search.semanticWeightWeight for semantic (vector) similarity (0-1)0.7
search.keywordWeightWeight for BM25 keyword matching (0-1)0.3
dashboard.enabledEnable the web dashboardtrue
dashboard.portPort for the dashboard server24300
dashboard.openBrowserAuto-open browser when dashboard startstrue
instructionsProject-specific instructions returned by get_project_instructionsNone

Default Behavior

Without a .lance-context.json file, lance-context will:

  • Index common source code files (TypeScript, JavaScript, Python, Go, Rust, Java, Ruby, PHP, C/C++, C#, Swift, Kotlin)
  • Exclude build artifacts, dependencies, and generated files
  • Use Gemini embeddings if GEMINI_API_KEY is set, otherwise use local Ollama with qwen3-embedding:0.6b
  • Split code into 100-line chunks with 20-line overlap
  • Use hybrid search with 70% semantic / 30% keyword weighting
  • Start the dashboard on port 24300

Environment Variables

Set these environment variables to configure embedding backends:

VariableDescriptionDefault
GEMINI_API_KEYGoogle Gemini API key for cloud embeddings (free tier available)None
OLLAMA_URLCustom Ollama server URL for local embeddingshttp://localhost:11434
LANCE_CONTEXT_PROJECTOverride the project path to indexCurrent working directory

Backend Selection Priority:

  • If embedding.backend is set in config, use that backend
  • If GEMINI_API_KEY is set, use Gemini
  • Fall back to Ollama (must be running locally)

Architecture

┌─────────────────────────────────────────────────────────────┐
│                    MCP Server (index.ts)                    │
│         Exposes tools: index_codebase, search_code          │
└─────────────────┬───────────────────────────────────────────┘
                  │
┌─────────────────▼───────────────────────────────────────────┐
│                  CodeIndexer (indexer.ts)                   │
│  - AST-aware chunking for supported languages               │
│  - Incremental indexing (only re-index changed files)       │
│  - Hybrid search (semantic + keyword scoring)               │
└─────────────────┬───────────────────────────────────────────┘
                  │
┌─────────────────▼───────────────────────────────────────────┐
│              Embedding Backends (embeddings/)               │
│            Gemini  │  Ollama (local)                        │
└─────────────────┬───────────────────────────────────────────┘
                  │
┌─────────────────▼───────────────────────────────────────────┐
│                   LanceDB Vector Store                      │
│           Stored in .lance-context/ directory               │
└─────────────────────────────────────────────────────────────┘

Embedding Backend Setup

lance-context automatically selects the best available backend (in priority order):

  • Google Gemini (if GEMINI_API_KEY is set, free tier available)

    export GEMINI_API_KEY=AIza...
    
  • Ollama (recommended for most users - free, local, no rate limits)

Ollama provides free, local embeddings with no API rate limits. Perfect for indexing large codebases.

Requirements: Ollama 0.2.0 or newer (for batch embedding API)

  • Install Ollama from ollama.com

  • Verify version (must be 0.2.0+):

    ollama --version
    
  • Pull the embedding model:

    ollama pull qwen3-embedding:0.6b
    
  • Verify it's working:

    ollama run qwen3-embedding:0.6b "test"
    

That's it! lance-context will automatically use Ollama when no Gemini API key is set.

Model Options

ModelSizeQualityBest For
qwen3-embedding:0.6b639MBGoodMost users (default)
qwen3-embedding:4b2.5GBBetterUsers with 16GB+ RAM
qwen3-embedding:8b4.7GBBestUsers with 32GB+ RAM

To use a different model, add to your .lance-context.json:

{
  "embedding": {
    "backend": "ollama",
    "model": "qwen3-embedding:4b"
  }
}

See Project Configuration for all configuration options including how to specify a backend.

Usage

Once installed, you'll have access to these tools:

index_codebase

Index your codebase for semantic search:

> index_codebase
Indexed 150 files, created 800 chunks.

With custom patterns:

> index_codebase(patterns: ["**/*.py"], excludePatterns: ["**/tests/**"])

search_code

Search using natural language:

> search_code(query: "authentication middleware")

## Result 1: src/middleware/auth.ts:1-50
...

get_index_status

Check index status:

> get_index_status
{
  "indexed": true,
  "fileCount": 150,
  "chunkCount": 800,
  "lastUpdated": "2024-12-27T12:00:00Z"
}

clear_index

Clear the index:

> clear_index
Index cleared.

get_project_instructions

Get project-specific instructions from the config:

> get_project_instructions
Use semantic search for exploring this codebase. Always run tests before committing.

Dashboard

lance-context includes a web dashboard for monitoring index status and usage.

Accessing the Dashboard

The dashboard starts automatically when the MCP server runs and is available at:

http://127.0.0.1:24300

The browser opens automatically on startup (configurable).

Dashboard Features

  • Index Status: Files indexed, chunks created, last updated time
  • Embedding Backend: Current backend and index path
  • Configuration: Project path, chunk settings, search weights
  • File Patterns: Include/exclude patterns being used
  • Command Usage: Real-time chart of MCP tool usage (using charts.css)
  • Beads Integration: Issue tracker status and ready tasks (if beads is configured)

Dashboard Configuration

Configure the dashboard via the dashboard options in .lance-context.json. See Configuration Options Reference for details.

How It Works

  • Indexing: Code files are chunked into ~100-line segments with overlap
  • Embedding: Each chunk is converted to a vector using your chosen backend
  • Storage: Vectors are stored in LanceDB (.lance-context/ directory)
  • Search: Natural language queries are embedded and matched against stored vectors

Supported Languages

TypeScript, JavaScript, Python, Go, Rust, Java, Ruby, PHP, C/C++, C#, Swift, Kotlin, and more.

Troubleshooting

"No embedding backend available"

This error means no API keys are set and Ollama is not running/accessible.

Solutions:

  • Set up Ollama (recommended):
    # Install from https://ollama.com, then:
    ollama pull qwen3-embedding:0.6b
    
  • Or set a Gemini API key: export GEMINI_API_KEY=AIza...

"Embedding dimension mismatch"

This occurs when switching between embedding backends (e.g., from Gemini to Ollama). Each backend produces different vector dimensions.

Solution: Force a full reindex:

> index_codebase(forceReindex: true)

Slow Indexing

Large codebases may take time to index initially.

Tips:

  • Use excludePatterns to skip unnecessary directories (tests, generated code)
  • Ollama is faster for local use but requires more resources
  • Subsequent runs use incremental indexing (only changed files)

Index Corruption

If you encounter strange search results or errors:

Solution: Clear and rebuild the index:

> clear_index
> index_codebase

Or manually delete the .lance-context/ directory and re-index.

License

MIT - See LICENSE for details.

Contributing

Contributions welcome! Please read our Contributing Guide before submitting PRs.

Credits

Built with:

Inspired by:

  • Serena by Oraios - Symbol-level code navigation and editing

Keywords

mcp

FAQs

Package last updated on 02 Feb 2026

Did you know?

Socket

Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.

Install

Related posts