black formatting

This commit is contained in:
Didier FOURNOUT 2025-02-13 16:12:46 +00:00
parent 6d62e71c34
commit 06062568c7
3 changed files with 10 additions and 8 deletions

View File

@ -115,7 +115,7 @@ async def send_post_request(
stream: bool = True, stream: bool = True,
key: Optional[str] = None, key: Optional[str] = None,
content_type: Optional[str] = None, content_type: Optional[str] = None,
user: UserModel = None user: UserModel = None,
): ):
r = None r = None
@ -296,7 +296,7 @@ async def update_config(
@cached(ttl=3) @cached(ttl=3)
async def get_all_models(request: Request, user: UserModel=None): async def get_all_models(request: Request, user: UserModel = None):
log.info("get_all_models()") log.info("get_all_models()")
if request.app.state.config.ENABLE_OLLAMA_API: if request.app.state.config.ENABLE_OLLAMA_API:
request_tasks = [] request_tasks = []
@ -317,7 +317,9 @@ async def get_all_models(request: Request, user: UserModel=None):
key = api_config.get("key", None) key = api_config.get("key", None)
if enable: if enable:
request_tasks.append(send_get_request(f"{url}/api/tags", key, user=user)) request_tasks.append(
send_get_request(f"{url}/api/tags", key, user=user)
)
else: else:
request_tasks.append(asyncio.ensure_future(asyncio.sleep(0, None))) request_tasks.append(asyncio.ensure_future(asyncio.sleep(0, None)))
@ -531,7 +533,7 @@ async def get_ollama_loaded_models(request: Request, user=Depends(get_verified_u
url, {} url, {}
), # Legacy support ), # Legacy support
).get("key", None), ).get("key", None),
user=user user=user,
) )
for idx, url in enumerate(request.app.state.config.OLLAMA_BASE_URLS) for idx, url in enumerate(request.app.state.config.OLLAMA_BASE_URLS)
] ]

View File

@ -52,7 +52,7 @@ log.setLevel(SRC_LOG_LEVELS["OPENAI"])
########################################## ##########################################
async def send_get_request(url, key=None, user: UserModel=None): async def send_get_request(url, key=None, user: UserModel = None):
timeout = aiohttp.ClientTimeout(total=AIOHTTP_CLIENT_TIMEOUT_OPENAI_MODEL_LIST) timeout = aiohttp.ClientTimeout(total=AIOHTTP_CLIENT_TIMEOUT_OPENAI_MODEL_LIST)
try: try:
async with aiohttp.ClientSession(timeout=timeout, trust_env=True) as session: async with aiohttp.ClientSession(timeout=timeout, trust_env=True) as session:
@ -70,7 +70,7 @@ async def send_get_request(url, key=None, user: UserModel=None):
if ENABLE_FORWARD_USER_INFO_HEADERS if ENABLE_FORWARD_USER_INFO_HEADERS
else {} else {}
), ),
} },
) as response: ) as response:
return await response.json() return await response.json()
except Exception as e: except Exception as e:

View File

@ -30,7 +30,7 @@ log = logging.getLogger(__name__)
log.setLevel(SRC_LOG_LEVELS["MAIN"]) log.setLevel(SRC_LOG_LEVELS["MAIN"])
async def get_all_base_models(request: Request, user: UserModel=None): async def get_all_base_models(request: Request, user: UserModel = None):
function_models = [] function_models = []
openai_models = [] openai_models = []
ollama_models = [] ollama_models = []
@ -59,7 +59,7 @@ async def get_all_base_models(request: Request, user: UserModel=None):
return models return models
async def get_all_models(request, user: UserModel=None): async def get_all_models(request, user: UserModel = None):
models = await get_all_base_models(request, user=user) models = await get_all_base_models(request, user=user)
# If there are no models, return an empty list # If there are no models, return an empty list