Report Abuse

Basic Information

Super-Rag is a developer-focused retrieval-augmented generation (RAG) pipeline for building AI applications that need scalable document ingestion, semantic search, and retrieval-backed question answering. The project provides a production-ready FastAPI REST API that accepts document ingestion requests, manages vector indexes, and serves query endpoints for retrieval and computation workflows. It supports multiple document formats, configurable splitting and chunking strategies, and pluggable encoders and vector databases. The code interpreter mode integrates with E2B.dev runtimes for computational Q&A. The README includes concrete API payload examples for ingest, query, and delete operations and instructions to run the service locally with a virtual environment and poetry. There is also an optional hosted Cloud API for quick starts. Overall the repo packages a full RAG pipeline with connectors and runtime options intended for developers deploying retrieval-backed AI features.

Links

Categorization

App Details

Features
The README highlights a set of production-oriented features: support for multiple document formats and a variety of vector databases including Pinecone, Qdrant, Weaviate, Astra, PGVector and planned Chroma. It exposes a REST API with endpoints for ingesting documents, querying indexes, and deleting entries. The pipeline offers customizable splitting and chunking parameters, configurable encoders (OpenAI, Cohere, HuggingFace, FastEmbed and planned providers), and encoder dimension settings. Interpreter mode enables running code interpreters for computational Q&A via E2B.dev custom runtimes. The system supports session management with session IDs for caching and micro-VM session persistence. The README includes payload examples showing encoder, splitter, and vector database configuration fields and optional filters for provider-specific filtering.
Use Cases
Super-Rag helps teams build retrieval-backed AI features by providing a ready-made pipeline and API that handle the common operational tasks of RAG systems. Developers can ingest URLs or files, configure document processing and embedding options, choose a vector store, and immediately query or delete documents via documented endpoints. The built-in interpreter mode enables richer Q&A that requires computation or code execution in sandboxes. Session IDs and caching support improve performance for repeated queries. The repository includes local installation and run instructions and a free Cloud API to get started quickly, reducing integration effort for projects that need semantic search, context retrieval, or retrieval-augmented generation capabilities.

Please fill the required fields*