nameOverride: "" ollama: # -- Automatically install Ollama Helm chart from https://otwld.github.io/ollama-helm/. Use [Helm Values](https://github.com/otwld/ollama-helm/#helm-values) to configure enabled: true # Example Ollama configuration with nvidia GPU enabled, automatically downloading a model, and deploying a PVC for model persistence # ollama: # gpu: # enabled: true # type: 'nvidia' # number: 1 # models: # - llama3 # runtimeClassName: nvidia # persistentVolume: # enabled: true # -- A list of Ollama API endpoints. These can be added in lieu of automatically installing the Ollama Helm chart, or in addition to it. ollamaUrls: [] # -- Value of cluster domain clusterDomain: cluster.local annotations: {} podAnnotations: {} replicaCount: 1 image: repository: ghcr.io/open-webui/open-webui tag: "" pullPolicy: Always resources: {} ingress: enabled: false class: "" # -- Use appropriate annotations for your Ingress controller, e.g., for NGINX: # nginx.ingress.kubernetes.io/rewrite-target: / annotations: {} host: "" tls: false existingSecret: "" persistence: enabled: true size: 2Gi existingClaim: "" # -- If using multiple replicas, you must update accessModes to ReadWriteMany accessModes: - ReadWriteOnce storageClass: "" selector: {} annotations: {} # -- Node labels for pod assignment. nodeSelector: {} # -- Tolerations for pod assignment tolerations: [] # -- Affinity for pod assignment affinity: {} service: type: ClusterIP annotations: {} port: 80 containerPort: 8080 nodePort: "" labels: {} loadBalancerClass: "" # -- Additional environments variables on the output Deployment definition. extraEnvVars: # - name: OPENAI_API_KEY # valueFrom: # secretKeyRef: # name: openai-api-key # key: api-key # - name: OLLAMA_DEBUG # value: "1"