From b7331433afd327fa9673c9538c96c2107752ce60 Mon Sep 17 00:00:00 2001 From: "Timothy J. Baek" Date: Wed, 22 May 2024 13:35:01 -0700 Subject: [PATCH] fix --- pipelines/examples/llama_cpp_pipeline.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/pipelines/examples/llama_cpp_pipeline.py b/pipelines/examples/llama_cpp_pipeline.py index c555993..47bf630 100644 --- a/pipelines/examples/llama_cpp_pipeline.py +++ b/pipelines/examples/llama_cpp_pipeline.py @@ -40,7 +40,7 @@ class Pipeline: print(body) response = self.llm.create_chat_completion_openai_v1( - messages=[message.model_dump() for message in messages], + messages=messages, stream=body["stream"], )