mirror of
https://github.com/open-webui/open-webui
synced 2024-11-23 00:27:40 +00:00
7.6 KiB
7.6 KiB
Changelog
All notable changes to this project will be documented in this file.
The format is based on Keep a Changelog, and this project adheres to Semantic Versioning.
[0.1.110] - 2024-03-06
Added
- 🌐 Multiple OpenAI Servers Support: Enjoy seamless integration with multiple OpenAI-compatible APIs, now supported natively.
Fixed
- 🔍 OCR Issue: Resolved PDF parsing issue caused by OCR malfunction.
- 🚫 RAG Issue: Fixed the RAG functionality, ensuring it operates smoothly.
- 📄 "Add Docs" Model Button: Addressed the non-functional behavior of the "Add Docs" model button.
[0.1.109] - 2024-03-06
Added
- 🔄 Multiple Ollama Servers Support: Enjoy enhanced scalability and performance with support for multiple Ollama servers in a single WebUI. Load balancing features are now available, providing improved efficiency (#788, #278).
- 🔧 Support for Claude 3 and Gemini: Responding to user requests, we've expanded our toolset to include Claude 3 and Gemini, offering a wider range of functionalities within our platform (#1064).
- 🔍 OCR Functionality for PDF Loader: We've augmented our PDF loader with Optical Character Recognition (OCR) capabilities. Now, extract text from scanned documents and images within PDFs, broadening the scope of content processing (#1050).
Fixed
- 🛠️ RAG Collection: Implemented a dynamic mechanism to recreate RAG collections, ensuring users have up-to-date and accurate data (#1031).
- 📝 User Agent Headers: Fixed issue of RAG web requests being sent with empty user_agent headers, reducing rejections from certain websites. Realistic headers are now utilized for these requests (#1024).
- ⏹️ Playground Cancel Functionality: Introducing a new "Cancel" option for stopping Ollama generation in the Playground, enhancing user control and usability (#1006).
- 🔤 Typographical Error in 'ASSISTANT' Field: Corrected a typographical error in the 'ASSISTANT' field within the GGUF model upload template for accuracy and consistency (#1061).
Changed
- 🔄 Refactored Message Deletion Logic: Streamlined message deletion process for improved efficiency and user experience, simplifying interactions within the platform (#1004).
- ⚠️ Deprecation of
OLLAMA_API_BASE_URL
: DeprecatedOLLAMA_API_BASE_URL
environment variable; recommend usingOLLAMA_BASE_URL
instead. Refer to our documentation for further details.
[0.1.108] - 2024-03-02
Added
- 🎮 Playground Feature (Beta): Explore the full potential of the raw API through an intuitive UI with our new playground feature, accessible to admins. Simply click on the bottom name area of the sidebar to access it. The playground feature offers two modes text completion (notebook) and chat completion. As it's in beta, please report any issues you encounter.
- 🛠️ Direct Database Download for Admins: Admins can now download the database directly from the WebUI via the admin settings.
- 🎨 Additional RAG Settings: Customize your RAG process with the ability to edit the TOP K value. Navigate to Documents > Settings > General to make changes.
- 🖥️ UI Improvements: Tooltips now available in the input area and sidebar handle. More tooltips will be added across other parts of the UI.
Fixed
- Resolved input autofocus issue on mobile when the sidebar is open, making it easier to use.
- Corrected numbered list display issue in Safari (#963).
- Restricted user ability to delete chats without proper permissions (#993).
Changed
- Simplified Ollama Settings: Ollama settings now don't require the
/api
suffix. You can now utilize the Ollama base URL directly, e.g.,http://localhost:11434
. Also, anOLLAMA_BASE_URL
environment variable has been added. - Database Renaming: Starting from this release,
ollama.db
will be automatically renamed towebui.db
.
[0.1.107] - 2024-03-01
Added
- 🚀 Makefile and LLM Update Script: Included Makefile and a script for LLM updates in the repository.
Fixed
- Corrected issue where links in the settings modal didn't appear clickable (#960).
- Fixed problem with web UI port not taking effect due to incorrect environment variable name in run-compose.sh (#996).
- Enhanced user experience by displaying chat in browser title and enabling automatic scrolling to the bottom (#992).
Changed
- Upgraded toast library from
svelte-french-toast
tosvelte-sonner
for a more polished UI. - Enhanced accessibility with the addition of dark mode on the authentication page.
[0.1.106] - 2024-02-27
Added
- 🎯 Auto-focus Feature: The input area now automatically focuses when initiating or opening a chat conversation.
Fixed
- Corrected typo from "HuggingFace" to "Hugging Face" (Issue #924).
- Resolved bug causing errors in chat completion API calls to OpenAI due to missing "num_ctx" parameter (Issue #927).
- Fixed issues preventing text editing, selection, and cursor retention in the input field (Issue #940).
- Fixed a bug where defining an OpenAI-compatible API server using 'OPENAI_API_BASE_URL' containing 'openai' string resulted in hiding models not containing 'gpt' string from the model menu. (Issue #930)
[0.1.105] - 2024-02-25
Added
- 📄 Document Selection: Now you can select and delete multiple documents at once for easier management.
Changed
- 🏷️ Document Pre-tagging: Simply click the "+" button at the top, enter tag names in the popup window, or select from a list of existing tags. Then, upload files with the added tags for streamlined organization.
[0.1.104] - 2024-02-25
Added
- 🔄 Check for Updates: Keep your system current by checking for updates conveniently located in Settings > About.
- 🗑️ Automatic Tag Deletion: Unused tags on the sidebar will now be deleted automatically with just a click.
Changed
- 🎨 Modernized Styling: Enjoy a refreshed look with updated styling for a more contemporary experience.
[0.1.103] - 2024-02-25
Added
-
🔗 Built-in LiteLLM Proxy: Now includes LiteLLM proxy within Open WebUI for enhanced functionality.
- Easily integrate existing LiteLLM configurations using
-v /path/to/config.yaml:/app/backend/data/litellm/config.yaml
flag. - When utilizing Docker container to run Open WebUI, ensure connections to localhost use
host.docker.internal
.
- Easily integrate existing LiteLLM configurations using
-
🖼️ Image Generation Enhancements: Introducing Advanced Settings with Image Preview Feature.
- Customize image generation by setting the number of steps; defaults to A1111 value.
Fixed
- Resolved issue with RAG scan halting document loading upon encountering unsupported MIME types or exceptions (Issue #866).
Changed
- Ollama is no longer required to run Open WebUI.
- Access our comprehensive documentation at Open WebUI Documentation.
[0.1.102] - 2024-02-22
Added
- 🖼️ Image Generation: Generate Images using the AUTOMATIC1111/stable-diffusion-webui API. You can set this up in Settings > Images.
- 📝 Change title generation prompt: Change the prompt used to generate titles for your chats. You can set this up in the Settings > Interface.
- 🤖 Change embedding model: Change the embedding model used to generate embeddings for your chats in the Dockerfile. Use any sentence transformer model from huggingface.co.
- 📢 CHANGELOG.md/Popup: This popup will show you the latest changes.
[0.1.101] - 2024-02-22
Fixed
- LaTex output formatting issue (#828)
Changed
- Instead of having the previous 1.0.0-alpha.101, we switched to semantic versioning as a way to respect global conventions.