mirror of
https://github.com/deepseek-ai/DeepSeek-Coder
synced 2024-12-04 18:14:44 +00:00
Update README.md
This commit is contained in:
parent
81c08522eb
commit
d3bb741e25
@ -131,7 +131,7 @@ model = AutoModelForCausalLM.from_pretrained("deepseek-ai/deepseek-coder-6.7b-in
|
|||||||
messages=[
|
messages=[
|
||||||
{ 'role': 'user', 'content': "write a quick sort algorithm in python."}
|
{ 'role': 'user', 'content': "write a quick sort algorithm in python."}
|
||||||
]
|
]
|
||||||
inputs = tokenizer.apply_chat_template(messages, return_tensors="pt").to(model.device)
|
inputs = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt").to(model.device)
|
||||||
# 32021 is the id of <|EOT|> token
|
# 32021 is the id of <|EOT|> token
|
||||||
outputs = model.generate(inputs, max_new_tokens=512, do_sample=False, top_k=50, top_p=0.95, num_return_sequences=1, eos_token_id=32021)
|
outputs = model.generate(inputs, max_new_tokens=512, do_sample=False, top_k=50, top_p=0.95, num_return_sequences=1, eos_token_id=32021)
|
||||||
print(tokenizer.decode(outputs[0][len(inputs[0]):], skip_special_tokens=True))
|
print(tokenizer.decode(outputs[0][len(inputs[0]):], skip_special_tokens=True))
|
||||||
|
Loading…
Reference in New Issue
Block a user