mirror of
https://github.com/open-webui/docs
synced 2025-05-19 18:58:41 +00:00
Include Helicone LLM observability with Open WebUI
This commit is contained in:
parent
a3d75146f0
commit
6d906bc3c8
@ -16,6 +16,16 @@ This tutorial is a community contribution and is not supported by the Open WebUI
|
|||||||
We're looking for talented individuals to create videos showcasing Open WebUI's features. If you make a video, we'll feature it at the top of our guide section!
|
We're looking for talented individuals to create videos showcasing Open WebUI's features. If you make a video, we'll feature it at the top of our guide section!
|
||||||
:::
|
:::
|
||||||
|
|
||||||
|
<iframe
|
||||||
|
width="560"
|
||||||
|
height="315"
|
||||||
|
src="https://www.youtube-nocookie.com/embed/8iVHOkUrpSA?si=Jt1GVqA0wY4UI7sF"
|
||||||
|
title="YouTube video player"
|
||||||
|
frameborder="0"
|
||||||
|
allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share"
|
||||||
|
allowfullscreen>
|
||||||
|
</iframe>
|
||||||
|
|
||||||
<iframe
|
<iframe
|
||||||
width="560"
|
width="560"
|
||||||
height="315"
|
height="315"
|
||||||
|
81
docs/tutorials/integrations/helicone.md
Normal file
81
docs/tutorials/integrations/helicone.md
Normal file
@ -0,0 +1,81 @@
|
|||||||
|
---
|
||||||
|
title: "🕵🏻♀️ Monitor your LLM requests with Helicone"
|
||||||
|
sidebar_position: 19
|
||||||
|
---
|
||||||
|
|
||||||
|
:::warning
|
||||||
|
This tutorial is a community contribution and is not supported by the Open WebUI team. It serves only as a demonstration on how to customize Open WebUI for your specific use case. Want to contribute? Check out the contributing tutorial.
|
||||||
|
:::
|
||||||
|
|
||||||
|
# Helicone Integration with Open WebUI
|
||||||
|
|
||||||
|
Helicone is an open-source LLM observability platform for developers to monitor, debug, and improve **production-ready** applications, including your Open WebUI deployment.
|
||||||
|
|
||||||
|
By enabling Helicone, you can log LLM requests, evaluate and experiment with prompts, and get instant insights that helps you push changes to production with confidence.
|
||||||
|
|
||||||
|
- **Real-time monitoring with consolidated view across model types**: Monitor both local Ollama models and cloud APIs through a single interface
|
||||||
|
- **Request visualization and replay**: See exactly what prompts were sent to each model in Open WebUI and the outputs generated by the LLMs for evaluation
|
||||||
|
- **Local LLM performance tracking**: Measure response times and throughput of your self-hosted models
|
||||||
|
- **Usage analytics by model**: Compare usage patterns between different models in your Open WebUI setup
|
||||||
|
- **User analytics** to understand interaction patterns
|
||||||
|
- **Debug capabilities** to troubleshoot issues with model responses
|
||||||
|
- **Cost tracking** for your LLM usage across providers
|
||||||
|
|
||||||
|
|
||||||
|
## How to integrate Helicone with OpenWebUI
|
||||||
|
|
||||||
|
<iframe
|
||||||
|
width="560"
|
||||||
|
height="315"
|
||||||
|
src="https://www.youtube-nocookie.com/embed/8iVHOkUrpSA?si=Jt1GVqA0wY4UI7sF"
|
||||||
|
title="YouTube video player"
|
||||||
|
frameborder="0"
|
||||||
|
allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share"
|
||||||
|
allowfullscreen>
|
||||||
|
</iframe>
|
||||||
|
|
||||||
|
### Step 1: Create a Helicone account and generate your API key
|
||||||
|
|
||||||
|
Create a [Helicone account](https://www.helicone.ai/) and log in to generate an [API key here](https://us.helicone.ai/settings/api-keys).
|
||||||
|
|
||||||
|
*— Make sure to generate a [write only API key](https://docs.helicone.ai/helicone-headers/helicone-auth). This ensures you only allow logging data to Helicone without read access to your private data.*
|
||||||
|
|
||||||
|
### Step 2: Create an OpenAI account and generate your API key
|
||||||
|
|
||||||
|
Create an OpenAI account and log into [OpenAI's Developer Portal](https://platform.openai.com/account/api-keys) to generate an API key.
|
||||||
|
|
||||||
|
### Step 3: Run your Open WebUI application using Helicone's base URL
|
||||||
|
|
||||||
|
To launch your first Open WebUI application, use the command from [Open WebUI docs](https://docs.openwebui.com/) and include Helicone's API BASE URL so you can query and monitor automatically.
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# Set your environment variables
|
||||||
|
export HELICONE_API_KEY=<YOUR_API_KEY>
|
||||||
|
export OPENAI_API_KEY=<YOUR_OPENAI_API_KEY>
|
||||||
|
|
||||||
|
# Run Open WebUI with Helicone integration
|
||||||
|
docker run -d -p 3000:8080 \
|
||||||
|
-e OPENAI_API_BASE_URL="https://oai.helicone.ai/v1/$HELICONE_API_KEY" \
|
||||||
|
-e OPENAI_API_KEY="$OPENAI_API_KEY" \
|
||||||
|
--name open-webui \
|
||||||
|
ghcr.io/open-webui/open-webui
|
||||||
|
```
|
||||||
|
|
||||||
|
If you already have a Open WebUI application deployed, go to the `Admin Panel` > `Settings` > `Connections` and click on the `+` sign for "Managing OpenAI API Connections". Update the following properties:
|
||||||
|
|
||||||
|
- Your `API Base URL` would be ``https://oai.helicone.ai/v1/<YOUR_HELICONE_API_KEY>``
|
||||||
|
- The `API KEY` would be your OpenAI API key.
|
||||||
|
|
||||||
|

|
||||||
|
|
||||||
|
### Step 4: Make sure monitoring is working
|
||||||
|
|
||||||
|
To make sure your integration is working, log into Helicone's dashboard and review the "Requests" tab.
|
||||||
|
|
||||||
|
You should see the requests you have made through your Open WebUI interface already being logged into Helicone.
|
||||||
|
|
||||||
|

|
||||||
|
|
||||||
|
## Learn more
|
||||||
|
|
||||||
|
For a comprehensive guide on Helicone, you can check out [Helicone's documentation here](https://docs.helicone.ai/getting-started/quick-start).
|
Loading…
Reference in New Issue
Block a user