mirror of
https://github.com/open-webui/open-webui
synced 2024-11-24 13:07:25 +00:00
Merge pull request #160 from ollama-webui/doc-update
doc: manual deployment instruction added
This commit is contained in:
commit
114ad7ecb8
61
README.md
61
README.md
@ -115,43 +115,53 @@ docker build -t ollama-webui .
|
|||||||
docker run -d -p 3000:8080 -e OLLAMA_API_BASE_URL=https://example.com/api --name ollama-webui --restart always ollama-webui
|
docker run -d -p 3000:8080 -e OLLAMA_API_BASE_URL=https://example.com/api --name ollama-webui --restart always ollama-webui
|
||||||
```
|
```
|
||||||
|
|
||||||
## How to Build for Static Deployment
|
## How to Install Without Docker
|
||||||
|
|
||||||
1. Clone & Enter the project
|
While we strongly recommend using our convenient Docker container installation for optimal support, we understand that some situations may require a non-Docker setup, especially for development purposes. Please note that non-Docker installations are not officially supported, and you might need to troubleshoot on your own.
|
||||||
|
|
||||||
|
### Project Components
|
||||||
|
|
||||||
|
The Ollama Web UI consists of two primary components: the frontend and the backend (which serves as a reverse proxy, handling static frontend files, and additional features). Both need to be running concurrently for the development environment using `npm run dev`. Alternatively, you can set the `PUBLIC_API_BASE_URL` during the build process to have the frontend connect directly to your Ollama instance or build the frontend as static files and serve them with the backend.
|
||||||
|
|
||||||
|
### Prerequisites
|
||||||
|
|
||||||
|
1. **Clone and Enter the Project:**
|
||||||
|
|
||||||
```sh
|
```sh
|
||||||
git clone https://github.com/ollama-webui/ollama-webui.git
|
git clone https://github.com/ollama-webui/ollama-webui.git
|
||||||
pushd ./ollama-webui/
|
cd ollama-webui/
|
||||||
```
|
```
|
||||||
|
|
||||||
2. Create and edit `.env`
|
2. **Create and Edit `.env`:**
|
||||||
|
|
||||||
```sh
|
```sh
|
||||||
cp -RPp example.env .env
|
cp -RPp example.env .env
|
||||||
```
|
```
|
||||||
|
|
||||||
3. Install node dependencies
|
### Building Ollama Web UI Frontend
|
||||||
|
|
||||||
|
1. **Install Node Dependencies:**
|
||||||
|
|
||||||
```sh
|
```sh
|
||||||
npm i
|
npm install
|
||||||
```
|
```
|
||||||
|
|
||||||
4. Run in dev mode, or build the site for deployment
|
2. **Run in Dev Mode or Build for Deployment:**
|
||||||
|
|
||||||
- Test in Dev mode:
|
- Dev Mode (requires the backend to be running simultaneously):
|
||||||
|
|
||||||
```sh
|
```sh
|
||||||
npm run dev
|
npm run dev
|
||||||
```
|
```
|
||||||
|
|
||||||
- Build for Deploy:
|
- Build for Deployment:
|
||||||
|
|
||||||
```sh
|
```sh
|
||||||
#`PUBLIC_API_BASE_URL` will overwrite the value in `.env`
|
# `PUBLIC_API_BASE_URL` overwrites the value in `.env`
|
||||||
PUBLIC_API_BASE_URL='https://example.com/api' npm run build
|
PUBLIC_API_BASE_URL='https://example.com/api' npm run build
|
||||||
```
|
```
|
||||||
|
|
||||||
5. Test the build with `caddy` (or the server of your choice)
|
3. **Test the Build with `Caddy` (or your preferred server):**
|
||||||
|
|
||||||
```sh
|
```sh
|
||||||
curl https://webi.sh/caddy | sh
|
curl https://webi.sh/caddy | sh
|
||||||
@ -160,6 +170,35 @@ docker run -d -p 3000:8080 -e OLLAMA_API_BASE_URL=https://example.com/api --name
|
|||||||
caddy run --envfile .env --config ./Caddyfile.localhost
|
caddy run --envfile .env --config ./Caddyfile.localhost
|
||||||
```
|
```
|
||||||
|
|
||||||
|
### Running Ollama Web UI Backend
|
||||||
|
|
||||||
|
If you wish to run the backend for deployment, ensure that the frontend is built so that the backend can serve the frontend files along with the API route.
|
||||||
|
|
||||||
|
#### Setup Instructions
|
||||||
|
|
||||||
|
1. **Install Python Requirements:**
|
||||||
|
|
||||||
|
```sh
|
||||||
|
cd ./backend
|
||||||
|
pip install -r requirements.txt
|
||||||
|
```
|
||||||
|
|
||||||
|
2. **Run Python Backend:**
|
||||||
|
|
||||||
|
- Dev Mode with Hot Reloading:
|
||||||
|
|
||||||
|
```sh
|
||||||
|
sh dev.sh
|
||||||
|
```
|
||||||
|
|
||||||
|
- Deployment:
|
||||||
|
|
||||||
|
```sh
|
||||||
|
sh start.sh
|
||||||
|
```
|
||||||
|
|
||||||
|
Now, you should have the Ollama Web UI up and running at [http://localhost:8080/](http://localhost:8080/). Feel free to explore the features and functionalities of Ollama! If you encounter any issues, please refer to the instructions above or reach out to the community for assistance.
|
||||||
|
|
||||||
## Troubleshooting
|
## Troubleshooting
|
||||||
|
|
||||||
See [TROUBLESHOOTING.md](/TROUBLESHOOTING.md) for information on how to troubleshoot and/or join our [Ollama Web UI Discord community](https://discord.gg/5rJgQTnV4s).
|
See [TROUBLESHOOTING.md](/TROUBLESHOOTING.md) for information on how to troubleshoot and/or join our [Ollama Web UI Discord community](https://discord.gg/5rJgQTnV4s).
|
||||||
|
Loading…
Reference in New Issue
Block a user