Merge pull request #160 from ollama-webui/doc-update

doc: manual deployment instruction added
This commit is contained in:
Timothy Jaeryang Baek 2023-11-29 16:11:24 -05:00 committed by GitHub
commit 114ad7ecb8
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23

View File

@ -115,43 +115,53 @@ docker build -t ollama-webui .
docker run -d -p 3000:8080 -e OLLAMA_API_BASE_URL=https://example.com/api --name ollama-webui --restart always ollama-webui docker run -d -p 3000:8080 -e OLLAMA_API_BASE_URL=https://example.com/api --name ollama-webui --restart always ollama-webui
``` ```
## How to Build for Static Deployment ## How to Install Without Docker
1. Clone & Enter the project While we strongly recommend using our convenient Docker container installation for optimal support, we understand that some situations may require a non-Docker setup, especially for development purposes. Please note that non-Docker installations are not officially supported, and you might need to troubleshoot on your own.
### Project Components
The Ollama Web UI consists of two primary components: the frontend and the backend (which serves as a reverse proxy, handling static frontend files, and additional features). Both need to be running concurrently for the development environment using `npm run dev`. Alternatively, you can set the `PUBLIC_API_BASE_URL` during the build process to have the frontend connect directly to your Ollama instance or build the frontend as static files and serve them with the backend.
### Prerequisites
1. **Clone and Enter the Project:**
```sh ```sh
git clone https://github.com/ollama-webui/ollama-webui.git git clone https://github.com/ollama-webui/ollama-webui.git
pushd ./ollama-webui/ cd ollama-webui/
``` ```
2. Create and edit `.env` 2. **Create and Edit `.env`:**
```sh ```sh
cp -RPp example.env .env cp -RPp example.env .env
``` ```
3. Install node dependencies ### Building Ollama Web UI Frontend
1. **Install Node Dependencies:**
```sh ```sh
npm i npm install
``` ```
4. Run in dev mode, or build the site for deployment 2. **Run in Dev Mode or Build for Deployment:**
- Test in Dev mode: - Dev Mode (requires the backend to be running simultaneously):
```sh ```sh
npm run dev npm run dev
``` ```
- Build for Deploy: - Build for Deployment:
```sh ```sh
#`PUBLIC_API_BASE_URL` will overwrite the value in `.env` # `PUBLIC_API_BASE_URL` overwrites the value in `.env`
PUBLIC_API_BASE_URL='https://example.com/api' npm run build PUBLIC_API_BASE_URL='https://example.com/api' npm run build
``` ```
5. Test the build with `caddy` (or the server of your choice) 3. **Test the Build with `Caddy` (or your preferred server):**
```sh ```sh
curl https://webi.sh/caddy | sh curl https://webi.sh/caddy | sh
@ -160,6 +170,35 @@ docker run -d -p 3000:8080 -e OLLAMA_API_BASE_URL=https://example.com/api --name
caddy run --envfile .env --config ./Caddyfile.localhost caddy run --envfile .env --config ./Caddyfile.localhost
``` ```
### Running Ollama Web UI Backend
If you wish to run the backend for deployment, ensure that the frontend is built so that the backend can serve the frontend files along with the API route.
#### Setup Instructions
1. **Install Python Requirements:**
```sh
cd ./backend
pip install -r requirements.txt
```
2. **Run Python Backend:**
- Dev Mode with Hot Reloading:
```sh
sh dev.sh
```
- Deployment:
```sh
sh start.sh
```
Now, you should have the Ollama Web UI up and running at [http://localhost:8080/](http://localhost:8080/). Feel free to explore the features and functionalities of Ollama! If you encounter any issues, please refer to the instructions above or reach out to the community for assistance.
## Troubleshooting ## Troubleshooting
See [TROUBLESHOOTING.md](/TROUBLESHOOTING.md) for information on how to troubleshoot and/or join our [Ollama Web UI Discord community](https://discord.gg/5rJgQTnV4s). See [TROUBLESHOOTING.md](/TROUBLESHOOTING.md) for information on how to troubleshoot and/or join our [Ollama Web UI Discord community](https://discord.gg/5rJgQTnV4s).