mirror of
https://github.com/clearml/clearml-docs
synced 2025-01-31 06:27:22 +00:00
Small edits
This commit is contained in:
parent
2cde779fad
commit
ec9c3da7ba
@ -3,7 +3,7 @@ title: vLLM Model Deployment
|
|||||||
---
|
---
|
||||||
|
|
||||||
:::important Enterprise Feature
|
:::important Enterprise Feature
|
||||||
The Model Deployment App is available under the ClearML Enterprise plan.
|
The vLLM Model Deployment App is available under the ClearML Enterprise plan.
|
||||||
:::
|
:::
|
||||||
|
|
||||||
The vLLM Model Deployment application enables users to quickly deploy LLM models as networking services over a secure
|
The vLLM Model Deployment application enables users to quickly deploy LLM models as networking services over a secure
|
||||||
|
@ -37,7 +37,7 @@ Applications for deploying user interfaces for models:
|
|||||||
### Deploy
|
### Deploy
|
||||||
Applications for deploying machine learning models as scalable, secure services:
|
Applications for deploying machine learning models as scalable, secure services:
|
||||||
* [**Embedding Model Deployment**](apps_embed_model_deployment.md) - Deploy embedding models as networking services over a secure endpoint (available under ClearML Enterprise Plan)
|
* [**Embedding Model Deployment**](apps_embed_model_deployment.md) - Deploy embedding models as networking services over a secure endpoint (available under ClearML Enterprise Plan)
|
||||||
* [**Model Deployment**](apps_model_deployment.md) - Deploy LLM models as networking services over a secure endpoint (available under ClearML Enterprise Plan)
|
* [**vLLM Model Deployment**](apps_model_deployment.md) - Deploy LLM models as networking services over a secure endpoint (available under ClearML Enterprise Plan)
|
||||||
* [**llama.cpp**](apps_llama_deployment.md) - Deploy LLM models in GGUF format using [`llama.cpp`](https://github.com/ggerganov/llama.cpp) as networking services over a secure endpoint (available under ClearML Enterprise Plan)
|
* [**llama.cpp**](apps_llama_deployment.md) - Deploy LLM models in GGUF format using [`llama.cpp`](https://github.com/ggerganov/llama.cpp) as networking services over a secure endpoint (available under ClearML Enterprise Plan)
|
||||||
|
|
||||||
:::info Autoscalers
|
:::info Autoscalers
|
||||||
|
Loading…
Reference in New Issue
Block a user