add some otel initialization, pointing at honeycomb + a function to wrap other functions with spans. use that function to wrap `callAnthropic` and add some context (number of messages sent, token usage, along with some metadata should we ever support other providers.)
The honeycomb api key/dataset are to be passed via env vars `HONEYCOMB_API_KEY` and `HONEYCOMB_DATASET` respectively.
* feat: Integrate AWS Bedrock with Claude 3.5 Sonnet, Claude 3 Sonnet, and Claude 3.5 Haiku
* update Dockerfile for AWS Bedrock configuration
* feat: add new Bedrock model 'Mistral' and update Haiku to version 3
* feat: add new bedrock model Nova Lite and Nova Pro
* Update README documentation to reflect the latest changes
* Add the icon for aws bedrock
* add support for serialized AWS Bedrock configuration in api key
- Create detailed provider implementation guide with:
- Architecture overview and implementation steps
- Configuration patterns and best practices
- Testing checklist and Docker integration guide
- Example using Together AI implementation
- Add Together AI as new provider with:
- Environment variables and Docker configuration
- Support for Qwen, Llama, and Mixtral models
- API key and base URL management
- OpenAI-compatible API integration