AgentRAG.io

Intelligent RAG with Autonomous Agents.

Build with

Current Features

  • Intelligent Document Search : ChromaDB-powered semantic search with adjustable retrieval parameters
  • Multi-Provider LLM Support : Seamless integration with OpenAI, Claude, Gemini, Ollama, and 10+ providers
  • Temperature Control : Per-conversation adjustment (0.0-2.0)
  • Project-Based Organization : Isolate document collections and conversations by project
  • Real-Time Streaming : Server-Sent Events for progressive responses
  • Fine-Grained Control : Per-conversation temperature, top-k, and context window management
  • Enterprise Security : JWT authentication, AES-256 encryption, GDPR compliance

Advanced Features Agent & RAG

  • RAG-Powered Context : Agents automatically retrieve relevant context from your document knowledge base
  • LLM Integration : Agents can call LLMs with intelligent prompts and context management
  • MCP Tool Access : Connect to external services (GitHub, Jira, Slack, linters, test runners, etc.)
  • Multi-Step Workflows : Orchestrate complex tasks (fetch repo → generate code → test → lint → commit → PR)
  • Real-Time Streaming : Live execution logs and progress updates via Server-Sent Events
  • Error Handling : Automatic retries, timeouts, and graceful degradation

Providers Supported

  • OpenAI : GPT-4o, GPT-4-turbo, o1-preview, o1-mini.
  • Anthropic Claude : Claude 3.5 Sonnet, Claude 3 Opus
  • Google Gemini : Gemini 1.5 Pro/Flash, Gemini 2.0
  • OpenRouter : 200+ models (free + paid)
  • xAI Grok : Grok-3, Grok-3-mini, Grok-3-vision
  • Groq : Mixtral, LLaMA 3, Gemma
  • HuggingFace : Zephyr, Mistral, LLaMA 2
  • Ollama : ollama pull llama3
  • LM Studio : GUI app
  • vLLM : Python + CUDA
  • LMDeploy : Python + TurboMind
  • Oobabooga : Web UI

About

This project is licensed under the MIT License - see the LICENSE file for details.

Contact me

Built with ❤️ for the LLM community

For questions, suggestions, or support, please open an issue or contact the maintainers.@me.