diff --git a/pipelines/examples/llama_cpp_pipeline.py b/pipelines/examples/llama_cpp_pipeline.py index c555993..47bf630 100644 --- a/pipelines/examples/llama_cpp_pipeline.py +++ b/pipelines/examples/llama_cpp_pipeline.py @@ -40,7 +40,7 @@ class Pipeline: print(body) response = self.llm.create_chat_completion_openai_v1( - messages=[message.model_dump() for message in messages], + messages=messages, stream=body["stream"], )