diff --git a/apps/llama/configs/config.yaml b/apps/llama/configs/config.yaml index 17a2f9d..59fb4a6 100644 --- a/apps/llama/configs/config.yaml +++ b/apps/llama/configs/config.yaml @@ -17,13 +17,6 @@ macros: thinking_on: "--chat-template-kwargs '{\"enable_thinking\": true}'" thinking_off: "--chat-template-kwargs '{\"enable_thinking\": false}'" -peers: - openrouter: - proxy: https://openrouter.ai/api - apiKey: ${env.OPENROUTER_API_KEY} - models: - - z-ai/glm-5 - hooks: on_startup: preload: diff --git a/apps/llama/deployment.yaml b/apps/llama/deployment.yaml index 44ccb76..7c4ada6 100644 --- a/apps/llama/deployment.yaml +++ b/apps/llama/deployment.yaml @@ -29,12 +29,6 @@ spec: - containerPort: 8080 name: http protocol: TCP - env: - - name: OPENROUTER_API_KEY - valueFrom: - secretKeyRef: - name: llama-openrouter - key: OPENROUTER_API_KEY volumeMounts: - name: models mountPath: /root/.cache diff --git a/apps/llama/secret.yaml b/apps/llama/secret.yaml index 82ab723..3809b61 100644 --- a/apps/llama/secret.yaml +++ b/apps/llama/secret.yaml @@ -36,26 +36,3 @@ spec: excludeRaw: true vaultAuthRef: llama ---- -apiVersion: secrets.hashicorp.com/v1beta1 -kind: VaultStaticSecret -metadata: - name: llama-openrouter - namespace: llama -spec: - type: kv-v2 - - mount: secret - path: openrouter - - destination: - create: true - name: llama-openrouter - type: Opaque - transformation: - excludeRaw: true - templates: - OPENROUTER_API_KEY: - text: '{{ get .Secrets "API_KEY" }}' - - vaultAuthRef: llama