diff --git a/docs/getting-started/index.md b/docs/getting-started/index.md
index 9d1c330..936a739 100644
--- a/docs/getting-started/index.md
+++ b/docs/getting-started/index.md
@@ -25,9 +25,11 @@ title: "🚀 Getting Started"
- Follow the installation instructions provided on the website. After installation, open Docker Desktop to ensure it's running properly.
#### For Ubuntu Users:
+
1. **Open your terminal.**
2. **Set up Docker's apt repository:**
+
- Update your package index:
```bash
sudo apt-get update
@@ -55,6 +57,7 @@ title: "🚀 Getting Started"
Note: If you're using an Ubuntu derivative distro, such as Linux Mint, you might need to use `UBUNTU_CODENAME` instead of `VERSION_CODENAME`.
3. **Install Docker Engine:**
+
- Update your package index again:
```bash
sudo apt-get update
@@ -85,78 +88,6 @@ title: "🚀 Getting Started"
-## One-line Command to Install Ollama and Open WebUI Together
-
-#### Using Docker Compose
-
-- If you don't have Ollama yet, use Docker Compose for easy installation. Run this command:
-
- ```bash
- docker compose up -d --build
- ```
-
-- **For GPU Support:** Use an additional Docker Compose file:
-
- ```bash
- docker compose -f docker-compose.yaml -f docker-compose.gpu.yaml up -d --build
- ```
-
-- **For AMD GPU Support:** Some AMD GPUs require setting an environment variable for proper functionality:
-
- ```bash
- HSA_OVERRIDE_GFX_VERSION=11.0.0 docker compose -f docker-compose.yaml -f docker-compose.amdgpu.yaml up -d --build
- ```
-
-
- AMD GPU Support with HSA_OVERRIDE_GFX_VERSION
-
- For AMD GPU users encountering compatibility issues, setting the `HSA_OVERRIDE_GFX_VERSION` environment variable is crucial. This variable instructs the ROCm platform to emulate a specific GPU architecture, ensuring compatibility with various AMD GPUs not officially supported. Depending on your GPU model, adjust the `HSA_OVERRIDE_GFX_VERSION` as follows:
-
- - **For RDNA1 & RDNA2 GPUs** (e.g., RX 6700, RX 680M): Use `HSA_OVERRIDE_GFX_VERSION=10.3.0`.
- - **For RDNA3 GPUs**: Set `HSA_OVERRIDE_GFX_VERSION=11.0.0`.
- - **For older GCN (Graphics Core Next) GPUs**: The version to use varies. GCN 4th gen and earlier might require different settings, such as `ROC_ENABLE_PRE_VEGA=1` for GCN4, or `HSA_OVERRIDE_GFX_VERSION=9.0.0` for Vega (GCN5.0) emulation.
-
- Ensure to replace `` with the appropriate version number based on your GPU model and the guidelines above. For a detailed list of compatible versions and more in-depth instructions, refer to the [ROCm documentation](https://rocm.docs.amd.com) and the [openSUSE Wiki on AMD GPGPU](https://en.opensuse.org/SDB:AMD_GPGPU).
-
- Example command for RDNA1 & RDNA2 GPUs:
- ```bash
- HSA_OVERRIDE_GFX_VERSION=10.3.0 docker compose -f docker-compose.yaml -f docker-compose.amdgpu.yaml up -d --build
- ```
-
-
-
-- **To Expose Ollama API:** Use another Docker Compose file:
-
- ```bash
- docker compose -f docker-compose.yaml -f docker-compose.api.yaml up -d --build
- ```
-
-#### Using `run-compose.sh` Script (Linux or Docker-Enabled WSL2 on Windows)
-
-- Give execute permission to the script:
-
- ```bash
- chmod +x run-compose.sh
- ```
-
-- For CPU-only container:
-
- ```bash
- ./run-compose.sh
- ```
-
-- For GPU support (read the note about GPU compatibility):
-
- ```bash
- ./run-compose.sh --enable-gpu
- ```
-
-- To build the latest local version, add `--build`:
-
- ```bash
- ./run-compose.sh --enable-gpu --build
- ```
-
## Quick Start with Docker 🐳
:::info
@@ -171,12 +102,26 @@ When using Docker to install Open WebUI, make sure to include the `-v open-webui
- **If Ollama is on a Different Server**, use this command:
-- To connect to Ollama on another server, change the `OLLAMA_BASE_URL` to the server's URL:
+ To connect to Ollama on another server, change the `OLLAMA_BASE_URL` to the server's URL:
```bash
docker run -d -p 3000:8080 -e OLLAMA_BASE_URL=https://example.com -v open-webui:/app/backend/data --name open-webui --restart always ghcr.io/open-webui/open-webui:main
```
+### Installing Both Open WebUI and Ollama Together
+
+- **With GPU Support**, Use this command:
+
+ ```bash
+ docker run -d -p 3000:8080 --gpus=all -v ollama:/root/.ollama -v open-webui:/app/backend/data --name open-webui --restart always ghcr.io/open-webui/open-webui:ollama
+ ```
+
+- **For CPU Only**, Use this command:
+
+ ```bash
+ docker run -d -p 3000:8080 -v ollama:/root/.ollama -v open-webui:/app/backend/data --name open-webui --restart always ghcr.io/open-webui/open-webui:ollama
+ ```
+
- After installation, you can access Open WebUI at [http://localhost:3000](http://localhost:3000). Enjoy! 😄
### GPU Support
@@ -228,6 +173,79 @@ docker run -d --network=host -v open-webui:/app/backend/data -e OLLAMA_BASE_URL=
For more details on networking in Docker and addressing common connectivity issues, visit our [FAQ page](/faq/). This page provides additional context and solutions for frequently encountered problems, ensuring a smoother operation of Open WebUI in various environments.
+## One-line Command to Install Ollama and Open WebUI Together
+
+#### Using Docker Compose
+
+- If you don't have Ollama yet, use Docker Compose for easy installation. Run this command:
+
+ ```bash
+ docker compose up -d --build
+ ```
+
+- **For GPU Support:** Use an additional Docker Compose file:
+
+ ```bash
+ docker compose -f docker-compose.yaml -f docker-compose.gpu.yaml up -d --build
+ ```
+
+- **For AMD GPU Support:** Some AMD GPUs require setting an environment variable for proper functionality:
+
+ ```bash
+ HSA_OVERRIDE_GFX_VERSION=11.0.0 docker compose -f docker-compose.yaml -f docker-compose.amdgpu.yaml up -d --build
+ ```
+
+
+ AMD GPU Support with HSA_OVERRIDE_GFX_VERSION
+
+ For AMD GPU users encountering compatibility issues, setting the `HSA_OVERRIDE_GFX_VERSION` environment variable is crucial. This variable instructs the ROCm platform to emulate a specific GPU architecture, ensuring compatibility with various AMD GPUs not officially supported. Depending on your GPU model, adjust the `HSA_OVERRIDE_GFX_VERSION` as follows:
+
+ - **For RDNA1 & RDNA2 GPUs** (e.g., RX 6700, RX 680M): Use `HSA_OVERRIDE_GFX_VERSION=10.3.0`.
+ - **For RDNA3 GPUs**: Set `HSA_OVERRIDE_GFX_VERSION=11.0.0`.
+ - **For older GCN (Graphics Core Next) GPUs**: The version to use varies. GCN 4th gen and earlier might require different settings, such as `ROC_ENABLE_PRE_VEGA=1` for GCN4, or `HSA_OVERRIDE_GFX_VERSION=9.0.0` for Vega (GCN5.0) emulation.
+
+ Ensure to replace `` with the appropriate version number based on your GPU model and the guidelines above. For a detailed list of compatible versions and more in-depth instructions, refer to the [ROCm documentation](https://rocm.docs.amd.com) and the [openSUSE Wiki on AMD GPGPU](https://en.opensuse.org/SDB:AMD_GPGPU).
+
+ Example command for RDNA1 & RDNA2 GPUs:
+
+ ```bash
+ HSA_OVERRIDE_GFX_VERSION=10.3.0 docker compose -f docker-compose.yaml -f docker-compose.amdgpu.yaml up -d --build
+ ```
+
+
+
+- **To Expose Ollama API:** Use another Docker Compose file:
+
+ ```bash
+ docker compose -f docker-compose.yaml -f docker-compose.api.yaml up -d --build
+ ```
+
+#### Using `run-compose.sh` Script (Linux or Docker-Enabled WSL2 on Windows)
+
+- Give execute permission to the script:
+
+ ```bash
+ chmod +x run-compose.sh
+ ```
+
+- For CPU-only container:
+
+ ```bash
+ ./run-compose.sh
+ ```
+
+- For GPU support (read the note about GPU compatibility):
+
+ ```bash
+ ./run-compose.sh --enable-gpu
+ ```
+
+- To build the latest local version, add `--build`:
+
+ ```bash
+ ./run-compose.sh --enable-gpu --build
+ ```
+
## Installing with Podman