mirror of
				https://github.com/open-webui/openapi-servers
				synced 2025-06-26 18:17:04 +00:00 
			
		
		
		
	| .. | ||
| summarizers | ||
| __init__.py | ||
| docker-compose.yml | ||
| main.py | ||
| README.md | ||
| requirements.txt | ||
📚 Local Summarizer Agent
This FastAPI server acts to summarize a given chunk of text.
It is assumed that you are running an ollama instance in an adjacent container with the default port available.
📦 Endpoints
POST /summarize/text
Summarizes the given block of text
📥 Request
Body:
{
    'text':'Your blob of text here. It can be unlimited, but is recommended to be within the context window of the LLM you are asking for a summary from.'
}
📤 Response:
{
    "status": "success",
    "summary": "A summary of your text."
}
POST /summarize/chat
Not yet implemented. Summarizes an exported Open WebUI chat JSON blob.
🧩 Environment Variables
| Name | Description | Default | 
|---|---|---|
| MODEL | The name of the model you are trying to reference. Should match the model in your ollama instance. | llama3 | 
| MODEL_URL | The URL path to the model you are trying to access. | http://host.docker.internal:11434 |