Autogen_GraphRAG_Ollama

Report Abuse

Basic Information

This repository provides an integrated reference implementation that combines GraphRAG, AutoGen agents, local Ollama LLMs, and Chainlit UI to build a fully local, multi-agent Retrieval-Augmented Generation (RAG) superbot. It demonstrates how to wire GraphRAG's knowledge search with AutoGen agent workflows using function calling, how to configure GraphRAG for both local and global search, and how to run inference and embeddings on offline Ollama models. The project includes practical setup steps for Linux and Windows such as installing Ollama models, creating a Python environment, initializing a GraphRAG root folder, replacing GraphRAG embedding modules with supplied utility files, generating embeddings and a knowledge graph, starting a Lite-LLM proxy server, and launching the Chainlit interactive UI. The repo is aimed at developers who want a reproducible, local RAG agent stack that avoids external OpenAI dependencies.

Links

App Details

Features
The codebase highlights several integration-focused features documented in the README. It implements Agentic-RAG by connecting GraphRAG search with AutoGen agents via function calling. It supports offline LLM inference and embeddings using local Ollama models such as mistral, nomic-embed-text, and llama3. It extends AutoGen to enable function calling with non-OpenAI LLMs through a Lite-LLM proxy server. The project supplies utility files to replace GraphRAG embedding modules, scripts to initialize and index a GraphRAG root, and instructions for creating embeddings and the knowledge graph. The user-facing component is a Chainlit UI that handles continuous conversations, multi-threading, and user input settings. Installation and run commands for both Linux and Windows are provided for reproducibility.
Use Cases
This repository helps developers and teams rapidly assemble a private, offline RAG system using locally hosted language models and agent orchestration. By documenting how to run Ollama models locally and route them through a Lite-LLM proxy, the project enables embedding creation and model inference without relying on external APIs. The integration with AutoGen showcases non-OpenAI function calling patterns, which is useful for teams wanting flexible agent behavior tied to local search and knowledge graphs. Chainlit UI support provides a ready interactive front end for sustained multi-turn conversations and testing multi-threaded agent flows. Cross-platform setup notes and concrete commands for initializing GraphRAG and building the knowledge graph reduce friction when reproducing the stack on Linux or Windows.

Please fill the required fields*