NeuronDB Documentation

GPU-accelerated vector search, model inference, hybrid retrieval, and RAG orchestration built into PostgreSQL. NeurondB is an AI PostgreSQL extension. Use this documentation to deploy NeurondB, operate background workers, and embed ML pipelines in SQL.

Documentation Library

Getting Started

Install NeurondB on PostgreSQL 16–18, verify GPU support, and apply baseline configuration.

Core Features

Learn how NeurondB models vectors, maintains indexes, and tunes recall versus latency.

ML & Embeddings

Generate, store, and serve embeddings alongside model lifecycle management.

  • Embeddings

    Transform text, audio, and images into dense vectors.

  • Inference

    Deploy ONNX models with GPU batching and caching.

  • Model Management

    Version control, approvals, and rollback workflows.

Hybrid Search & Reranking

Combine text search, BM25, and neural rerankers for production retrieval pipelines.

Background Workers

Operational guidance for queue execution, auto-tuning, and index maintenance workers.

Components

NeuronDB ecosystem components: NeuronAgent for agent runtime and NeuronMCP for MCP protocol support.

  • NeuronAgent

    AI agent runtime with REST API, WebSocket, long-term memory, and tool execution.

  • NeuronMCP

    Model Context Protocol server for MCP-compatible clients like Claude Desktop.