⭐ Features
Key Features of Open WebUI ⭐
-
🚀 Effortless Setup: Install seamlessly using Docker, Kubernetes, Podman, Helm Charts (
kubectl,kustomize,podman, orhelm) for a hassle-free experience with support for both:ollamaimage with bundled Ollama and:cudawith CUDA support. Learn more in our Quick Start Guide. -
🛠️ Guided Initial Setup: Complete the setup process with clarity, including an explicit indication of creating an admin account during the first-time setup.
-
🤝 Universal API Compatibility: Effortlessly integrate with any backend that follows the OpenAI Chat Completions protocol. This includes official OpenAI endpoints alongside dozens of third-party and local providers. The API URL can be customized to integrate Open WebUI seamlessly into your existing infrastructure. See Setup Guide.
-
🛡️ Granular Permissions and User Groups: By allowing administrators to create detailed user roles, user groups, and permissions across the workspace, we ensure a secure user environment for all users involved. This granularity not only enhances security, but also allows for customized user experiences, fostering a sense of ownership and responsibility amongst users. Learn more about RBAC.
-
🔐 SCIM 2.0 Provisioning: Enterprise-grade user and group provisioning through SCIM 2.0 protocol, enabling seamless integration with identity providers like Okta, Azure AD, and Google Workspace for automated user lifecycle management. Read the SCIM Guide.
-
📂 Centralized File Management: A unified dashboard to search, view, and manage all your uploaded documents in one place. Includes automated cleanup of Knowledge Base associations and vector embeddings when deleting files. Learn about File Management.
-
💬 Shared Chat Management: A centralized interface to audit every conversation you've ever shared. Easily search through your shared history, re-copy links, or revoke (unshare) access instantly from a single location. Learn about Shared Chats.
-
📱 Responsive Design: Enjoy a seamless experience across desktop PCs, laptops, and mobile devices.
-
📱 Progressive Web App for Mobile: Enjoy a native progressive web application experience on your mobile device with offline access on
localhostor a personal domain, and a smooth user interface. In order for our PWA to be installable on your device, it must be delivered in a secure context. This usually means that it must be served over HTTPS.info- To set up a PWA, you'll need some understanding of technologies like Linux, Docker, and reverse proxies such as
Nginx,Caddy, orTraefik. Using these tools can help streamline the process of building and deploying a PWA tailored to your needs. While there's no "one-click install" option available, and your available option to securely deploy your Open WebUI instance over HTTPS requires user experience, using these resources can make it easier to create and deploy a PWA tailored to your needs.
- To set up a PWA, you'll need some understanding of technologies like Linux, Docker, and reverse proxies such as
-
✒️🔢 Full Markdown and LaTeX Support: Elevate your LLM experience with comprehensive Markdown, LaTex, and Rich Text capabilities for enriched interaction. Explore Interface Features.
-
🧩 Model Builder: Easily create custom models from base Ollama models directly from Open WebUI. Create and add custom characters and agents, customize model elements, and import models effortlessly through Open WebUI Community integration. Learn more about Models.
-
📚 Advanced RAG Integration with Multiple Vector Databases: Dive into the future of chat interactions with cutting-edge Retrieval Augmented Generation (RAG) technology. Choose from 9 vector database options: ChromaDB (default), PostgreSQL with PGVector, Qdrant, Milvus, Elasticsearch, OpenSearch, Pinecone, S3Vector, and Oracle 23ai. Documents can be loaded into the
Documentstab of the Workspace and accessed using the pound key [#] before a query, or by starting the prompt with [#] followed by a URL for webpage content integration. Learn more about RAG. -
📄 Advanced Document Extraction with Multiple Engines: Extract text and data from various document formats including PDFs, Word documents, Excel spreadsheets, PowerPoint presentations, and more using your choice of extraction engines: Apache Tika, Docling, Azure Document Intelligence, Mistral OCR, or external custom (self-built) content extraction engines/document loaders. Advanced document processing capabilities enable seamless integration with your knowledge base, preserving structure and formatting while supporting OCR for scanned documents and images. Read about Document Extraction.
-
🔍 Web Search for RAG & Agentic Research: Perform web searches using 15+ providers including SearXNG, Google PSE, Brave Search, Kagi, Mojeek, Bocha, Tavily, Perplexity, and more. When using Native Function Calling, models can perform multiple searches sequentially and use the
fetch_urltool to read full page content for deep research. Learn about Agentic Search. -
🌐 Web Browsing & URL Fetching: Integrate websites by using the
#command or allow the model to independently visit links using thefetch_urltool in Native Mode, extracting full text content for precise analysis. -
🎨 Image Generation & Editing Integration: Seamlessly create and edit images using engines like DALL-E, Gemini, ComfyUI, and AUTOMATIC1111. Supports Native Tool Calling, allowing models to independently generate and refine images during a conversation. Learn more about Image Gen.
-
⚙️ Concurrent Model Utilization: Effortlessly engage with multiple models simultaneously, harnessing their unique strengths for optimal responses. Leverage a diverse set of model modalities in parallel to enhance your experience.
-
🔐 Role-Based Access Control (RBAC): Ensure secure access with restricted permissions. Only authorized individuals can access your Ollama, while model creation and pulling rights are exclusively reserved for administrators. Learn more about RBAC.
-
🌐🌍 Multilingual Support: Experience Open WebUI in your preferred language with our internationalization (
i18n) support. We invite you to join us in expanding our supported languages! We're actively seeking contributors! -
💾 Persistent Artifact Storage: Built-in key-value storage API for artifacts, enabling features like journals, trackers, leaderboards, and collaborative tools with both personal and shared data scopes that persist across sessions. Explore Chat Features.
-
☁️ Cloud Storage Integration: Native support for cloud storage backends including Amazon S3 (with S3-compatible providers), Google Cloud Storage, and Microsoft Azure Blob Storage for scalable file storage and data management. See Storage Config.
-
☁️ Enterprise Cloud Integration: Seamlessly import documents from Google Drive and OneDrive/SharePoint directly through the file picker interface, enabling smooth workflows with enterprise cloud storage solutions. Learn more in Environment Config and check out the SharePoint Guide.
-
📊 Production Observability with OpenTelemetry: Built-in OpenTelemetry support for comprehensive monitoring with traces, metrics, and logs export to your existing observability stack (Prometheus, Grafana, Jaeger, etc.), enabling production-grade monitoring and debugging. See Observability Config.
-
🔒 Encrypted Database Support: Optional at-rest encryption for SQLite databases using SQLCipher, providing enhanced security for sensitive data in smaller deployments without requiring PostgreSQL infrastructure. See Database Encryption.
-
⚖️ Horizontal Scalability for Production: Redis-backed session management and WebSocket support enabling multi-worker and multi-node deployments behind load balancers for high-availability production environments. See Advanced Topics and our Multi-Replica Guide.
-
🌟 Continuous Updates: We are committed to improving Open WebUI with regular updates, fixes, and new features.
And many more remarkable features including... ⚡️
🔧 Pipelines Support
-
🔧 Pipelines Framework: Seamlessly integrate and customize your Open WebUI experience with our modular plugin framework for enhanced customization and functionality (https://github.com/open-webui/pipelines). Our framework allows for the easy addition of custom logic and integration of Python libraries, from AI agents to home automation APIs. Perfect for plugin and tool development, as well as creating custom functions and filters. Learn more about Pipelines.
-
�️ Native Python Function Calling: Access the power of Python directly within Open WebUI with native function calling. Easily integrate custom code to build unique features like custom RAG pipelines, web search tools, and even agent-like actions via a built-in code editor to seamlessly develop and integrate function code within the
ToolsandFunctionsworkspace. Learn more about Tools. -
�📥 Upload Pipeline: Pipelines can be uploaded directly from the
Admin Panel>Settings>Pipelinesmenu, streamlining the pipeline management process.
The possibilities with our Pipelines framework knows no bounds and are practically limitless. Start with a few pre-built pipelines to help you get started!
-
🔗 Function Calling: Integrate Function Calling seamlessly through Pipelines to enhance your LLM interactions with advanced function calling capabilities.
-
📚 Custom RAG: Integrate a custom Retrieval Augmented Generation (RAG) pipeline seamlessly to enhance your LLM interactions with custom RAG logic.
-
📊 Message Monitoring with Langfuse: Monitor and analyze message interactions in real-time usage statistics via Langfuse pipeline.
-
⚖️ User Rate Limiting: Manage API usage efficiently by controlling the flow of requests sent to LLMs to prevent exceeding rate limits with Rate Limit pipeline.
-
🌍 Real-Time LibreTranslate Translation: Integrate real-time translations into your LLM interactions using LibreTranslate pipeline, enabling cross-lingual communication.
- Please note that this pipeline requires further setup with LibreTranslate in a Docker container to work.
-
🛡️ Toxic Message Filtering: Our Detoxify pipeline automatically filters out toxic messages to maintain a clean and safe chat environment.
-
🔒 LLM-Guard: Ensure secure LLM interactions with LLM-Guard pipeline, featuring a Prompt Injection Scanner that detects and mitigates crafty input manipulations targeting large language models. This protects your LLMs from data leakage and adds a layer of resistance against prompt injection attacks.
-
🕒 Conversation Turn Limits: Improve interaction management by setting limits on conversation turns with Conversation Turn Limit pipeline.
-
📈 OpenAI Generation Stats: Our OpenAI pipeline provides detailed generation statistics for OpenAI models.
-
🚀 Multi-Model Support: Our seamless integration with various AI models from various providers expands your possibilities with a wide range of language models to select from and interact with.
In addition to the extensive features and customization options, we also provide a library of example pipelines ready to use along with a practical example scaffold pipeline to help you get started. These resources will streamline your development process and enable you to quickly create powerful LLM interactions using Pipelines and Python. Happy coding! 💡
🖥️ User Experience
-
🖥️ Intuitive Interface: The chat interface has been designed with the user in mind, drawing inspiration from the user interface of ChatGPT.
-
⚡ Swift Responsiveness: Enjoy reliably fast and responsive performance.
-
🎨 Splash Screen: A simple loading splash screen for a smoother user experience.
-
🌐 Personalized Interface: Choose between a freshly designed search landing page and the classic chat UI from Settings > Interface, allowing for a tailored experience. Explore Interface Options.