From b5a433ae1b37f12821137308b6dc6152682954be Mon Sep 17 00:00:00 2001 From: BingxuanWang <151806429+BingxuanWang@users.noreply.github.com> Date: Tue, 28 Nov 2023 15:07:01 +0800 Subject: [PATCH] Update README.md --- README.md | 10 ++++++++-- 1 file changed, 8 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index d0241a5..130f466 100644 --- a/README.md +++ b/README.md @@ -327,9 +327,13 @@ The reproducible code for the following evaluation results can be found in the [ ### 7. Q&A -#### Could You Provide the tokenizer.model File for GGUF Model Quantization? +#### Could You Provide the tokenizer.model File for Model Quantization? -DeepSeek Coder utilizes the [HuggingFace Tokenizer](https://huggingface.co/docs/tokenizers/index) to implement the Bytelevel-BPE algorithm, with specially designed pre-tokenizers to ensure optimal performance. Currently, there is no direct way to convert the tokenizer into a SentencePiece tokenizer. We have submitted a [PR](https://github.com/ggerganov/llama.cpp/pull/4070) to the popular quantization repository [llama.cpp](https://github.com/ggerganov/llama.cpp) to fully support all HuggingFace pre-tokenizers, including ours. +DeepSeek Coder utilizes the [HuggingFace Tokenizer](https://huggingface.co/docs/tokenizers/index) to implement the Bytelevel-BPE algorithm, with specially designed pre-tokenizers to ensure optimal performance. Currently, there is no direct way to convert the tokenizer into a SentencePiece tokenizer. We are contributing to the open-source quantization methods facilitate the usage of HuggingFace Tokenizer. + +##### GGUF(llama.cpp) + +We have submitted a [PR](https://github.com/ggerganov/llama.cpp/pull/4070) to the popular quantization repository [llama.cpp](https://github.com/ggerganov/llama.cpp) to fully support all HuggingFace pre-tokenizers, including ours. While waiting for the PR to be merged, you can generate your GGUF model using the following steps: @@ -346,7 +350,9 @@ python convert-hf-to-gguf.py --outfile --model-name dee ./quantize q4_0 ./main -m -n 128 -p ``` +##### GPTQ(exllamav2) +The [PR](https://github.com/turboderp/exllamav2/pull/189) to [exllamav2](https://github.com/turboderp/exllamav2) is also under reviewing. Please using the forked version before the PR getting merged. ### 8. Resources [awesome-deepseek-coder](https://github.com/deepseek-ai/awesome-deepseek-coder) is a curated list of open-source projects related to DeepSeek Coder.