diff --git a/docs/getting-started/using-openwebui/OllamaModels.mdx b/docs/getting-started/using-openwebui/OllamaModels.mdx new file mode 100644 index 0000000..34f0244 --- /dev/null +++ b/docs/getting-started/using-openwebui/OllamaModels.mdx @@ -0,0 +1,81 @@ +--- +title: "🤖Ollama Models" +--- + +import Tabs from '@theme/Tabs'; +import TabItem from '@theme/TabItem'; + +# Ollama Models + +Explore how to download, load, and use models with Ollama, both via **Docker** and **Remote** setups. + +--- + + + + ## 🐳 Ollama Inside Docker + + If **Ollama is deployed inside Docker** (e.g., using Docker Compose or Kubernetes), the service will be available: + + - **Inside the container**: `http://127.0.0.1:11434` + - **From the host**: `http://localhost:11435` (if exposed via host network) + + ### Step 1: Check Available Models + ```bash + docker exec -it openwebui curl http://ollama:11434/v1/models + ``` + + From the host (if exposed): + ```bash + curl http://localhost:11435/v1/models + ``` + + ### Step 2: Download Llama 3.2 + ```bash + docker exec -it ollama ollama pull llama3.2 + ``` + + You can also download a higher-quality version (8-bit) from Hugging Face: + ```bash + docker exec -it ollama ollama pull hf.co/bartowski/Llama-3.2-3B-Instruct-GGUF:Q8_0 + ``` + + + + + ## 🛠️ Bring Your Own Ollama (BYO Ollama) + + If Ollama is running on the **host machine** or another server on your network, follow these steps. + + ### Step 1: Check Available Models + Local: + ```bash + curl http://localhost:11434/v1/models + ``` + + Remote: + ```bash + curl http://:11434/v1/models + ``` + + ### Step 2: Set the OLLAMA_BASE_URL + ```bash + export OLLAMA_HOST=:11434 + ``` + + ### Step 3: Download Llama 3.2 + ```bash + ollama pull llama3.2 + ``` + + Or download the 8-bit version from Hugging Face: + ```bash + ollama pull hf.co/bartowski/Llama-3.2-3B-Instruct-GGUF:Q8_0 + ``` + + + + +--- + +You now have everything you need to download and run models with **Ollama**. Happy exploring!