mirror of
https://github.com/open-webui/docs
synced 2025-05-20 19:26:22 +00:00
1.0 KiB
1.0 KiB
🛠️ Bring Your Own Ollama (BYO Ollama)
If Ollama is running on the host machine or another server on your network, follow these steps.
Step 1: Check Available Models
-
If Ollama is local, run:
curl http://localhost:11434/v1/models
-
If Ollama is remote, use:
curl http://<remote-ip>:11434/v1/models
This confirms that Ollama is available and lists the available models.
Step 2: Set the OLLAMA_BASE_URL
If Ollama is running remotely or on the host, set the following environment variable:
export OLLAMA_HOST=<remote-ip>:11434
This ensures Open WebUI can reach the remote Ollama instance.
Step 3: Download Llama 3.2
From your local or remote machine, run:
ollama pull llama3.2
Tip: Use this command to download the 8-bit version from Hugging Face:
ollama pull hf.co/bartowski/Llama-3.2-3B-Instruct-GGUF:Q8_0
Step 4: Access the WebUI
You can now access the WebUI at: