From f6948c24a8e123c74487fac147557e281cc5c930 Mon Sep 17 00:00:00 2001 From: Yineng Zhang Date: Mon, 20 Jan 2025 20:30:16 +0800 Subject: [PATCH] docs: serving DeepSeek-R1-Distill models with SGLang --- README.md | 6 ++++++ 1 file changed, 6 insertions(+) diff --git a/README.md b/README.md index e5bcaca..6e80244 100644 --- a/README.md +++ b/README.md @@ -191,6 +191,12 @@ For instance, you can easily start a service using [vLLM](https://github.com/vll vllm serve deepseek-ai/DeepSeek-R1-Distill-Qwen-32B --tensor-parallel-size 2 --max-model-len 32768 --enforce-eager ``` +You can also easily start a service using [SGLang](https://github.com/sgl-project/sglang) + +```bash +python3 -m sglang.launch_server --model deepseek-ai/DeepSeek-R1-Distill-Qwen-32B --trust-remote-code --tp 2 +``` + **NOTE: We recommend setting an appropriate temperature (between 0.5 and 0.7) when running these models, otherwise you may encounter issues with endless repetition or incoherent output.** ## 7. License