LLM.io

Multi-Provider Chat Application.

Build with

Current Features

  • Multi-Provider : 13+ providers with automatic fallback
  • SSE Streaming : Real-time responses with auto-reconnect
  • Temperature Control : Per-conversation adjustment (0.0-2.0)
  • Prompt Templates : Create, share, use {{placeholders}}
  • Token Analytics : Dashboard with daily/provider/model totals
  • Conversation History : PostgreSQL storage with full-text search
  • Export Conversations : Markdown, JSON, HTML

Advanced Features

  • Chain of Thought : Auto-detection for reasoning models (o1-preview, DeepSeek-R1, etc.)
  • Deep Reasoning Mode : Enables high max_tokens for long reasoning
  • Model Auto-discovery : Dynamically fetches available models per provider
  • Health Checks : Connection testing with latency monitoring
  • Rate Limiting : Abuse protection (60 req/min, configurable)
  • Multi-origin CORS : Simultaneous frontend/mobile/desktop support

Providers Supported

  • OpenAI : GPT-4o, GPT-4-turbo, o1-preview, o1-mini.
  • Anthropic Claude : Claude 3.5 Sonnet, Claude 3 Opus
  • Google Gemini : Gemini 1.5 Pro/Flash, Gemini 2.0
  • OpenRouter : 200+ models (free + paid)
  • xAI Grok : Grok-3, Grok-3-mini, Grok-3-vision
  • Groq : Mixtral, LLaMA 3, Gemma
  • HuggingFace : Zephyr, Mistral, LLaMA 2
  • Ollama : ollama pull llama3
  • LM Studio : GUI app
  • vLLM : Python + CUDA
  • LMDeploy : Python + TurboMind
  • Oobabooga : Web UI

About

This project is licensed under the MIT License - see the LICENSE file for details.

Contact me

Built with ❤️ for the LLM community

For questions, suggestions, or support, please open an issue or contact the maintainers.@me.