mirror of
https://github.com/open-webui/pipelines
synced 2025-06-26 18:15:58 +00:00
Refactor payload structure to comply with Bedrock Converse API
- Updated `inferenceConfig` to include `temperature`, `topP`, `maxTokens`, and `stopSequences`. - Added `additionalModelRequestFields` with `top_k` parameter. - Removed unnecessary deletion of `additionalModelRequestFields` in `stream_response` method.
This commit is contained in:
parent
827b47d2d5
commit
51e267c10f
@ -140,8 +140,13 @@ class Pipeline:
|
|||||||
payload = {"modelId": model_id,
|
payload = {"modelId": model_id,
|
||||||
"messages": processed_messages,
|
"messages": processed_messages,
|
||||||
"system": [{'text': system_message["content"] if system_message else 'you are an intelligent ai assistant'}],
|
"system": [{'text': system_message["content"] if system_message else 'you are an intelligent ai assistant'}],
|
||||||
"inferenceConfig": {"temperature": body.get("temperature", 0.5)},
|
"inferenceConfig": {
|
||||||
"additionalModelRequestFields": {"top_k": body.get("top_k", 200), "top_p": body.get("top_p", 0.9)}
|
"temperature": body.get("temperature", 0.5),
|
||||||
|
"topP": body.get("top_p", 0.9),
|
||||||
|
"maxTokens": body.get("max_tokens", 4096),
|
||||||
|
"stopSequences": body.get("stop", []),
|
||||||
|
},
|
||||||
|
"additionalModelRequestFields": {"top_k": body.get("top_k", 200)}
|
||||||
}
|
}
|
||||||
if body.get("stream", False):
|
if body.get("stream", False):
|
||||||
return self.stream_response(model_id, payload)
|
return self.stream_response(model_id, payload)
|
||||||
@ -166,8 +171,6 @@ class Pipeline:
|
|||||||
}
|
}
|
||||||
|
|
||||||
def stream_response(self, model_id: str, payload: dict) -> Generator:
|
def stream_response(self, model_id: str, payload: dict) -> Generator:
|
||||||
if "additionalModelRequestFields" in payload:
|
|
||||||
del payload["additionalModelRequestFields"]
|
|
||||||
streaming_response = self.bedrock_runtime.converse_stream(**payload)
|
streaming_response = self.bedrock_runtime.converse_stream(**payload)
|
||||||
for chunk in streaming_response["stream"]:
|
for chunk in streaming_response["stream"]:
|
||||||
if "contentBlockDelta" in chunk:
|
if "contentBlockDelta" in chunk:
|
||||||
|
Loading…
Reference in New Issue
Block a user