diff --git a/backend/open_webui/apps/audio/main.py b/backend/open_webui/apps/audio/main.py index 4734b0d95..345e0390a 100644 --- a/backend/open_webui/apps/audio/main.py +++ b/backend/open_webui/apps/audio/main.py @@ -443,7 +443,7 @@ def get_available_models() -> list[dict]: try: response = requests.get( - "https://api.elevenlabs.io/v1/models", headers=headers + "https://api.elevenlabs.io/v1/models", headers=headers, timeout=5 ) response.raise_for_status() models = response.json() diff --git a/src/lib/components/admin/Settings/Connections.svelte b/src/lib/components/admin/Settings/Connections.svelte index fe71e4816..97a760110 100644 --- a/src/lib/components/admin/Settings/Connections.svelte +++ b/src/lib/components/admin/Settings/Connections.svelte @@ -150,18 +150,20 @@ })() ]); - OPENAI_API_BASE_URLS.forEach(async (url, idx) => { - const res = await getOpenAIModels(localStorage.token, idx); - if (res.pipelines) { - pipelineUrls[url] = true; - } - }); - const ollamaConfig = await getOllamaConfig(localStorage.token); const openaiConfig = await getOpenAIConfig(localStorage.token); ENABLE_OPENAI_API = openaiConfig.ENABLE_OPENAI_API; ENABLE_OLLAMA_API = ollamaConfig.ENABLE_OLLAMA_API; + + if (ENABLE_OPENAI_API) { + OPENAI_API_BASE_URLS.forEach(async (url, idx) => { + const res = await getOpenAIModels(localStorage.token, idx); + if (res.pipelines) { + pipelineUrls[url] = true; + } + }); + } } });