get rid of openrouter proxying via llama-swap
All checks were successful
ci/woodpecker/push/flux-reconcile-source Pipeline was successful
All checks were successful
ci/woodpecker/push/flux-reconcile-source Pipeline was successful
This commit is contained in:
@@ -17,13 +17,6 @@ macros:
|
|||||||
thinking_on: "--chat-template-kwargs '{\"enable_thinking\": true}'"
|
thinking_on: "--chat-template-kwargs '{\"enable_thinking\": true}'"
|
||||||
thinking_off: "--chat-template-kwargs '{\"enable_thinking\": false}'"
|
thinking_off: "--chat-template-kwargs '{\"enable_thinking\": false}'"
|
||||||
|
|
||||||
peers:
|
|
||||||
openrouter:
|
|
||||||
proxy: https://openrouter.ai/api
|
|
||||||
apiKey: ${env.OPENROUTER_API_KEY}
|
|
||||||
models:
|
|
||||||
- z-ai/glm-5
|
|
||||||
|
|
||||||
hooks:
|
hooks:
|
||||||
on_startup:
|
on_startup:
|
||||||
preload:
|
preload:
|
||||||
|
|||||||
@@ -36,26 +36,3 @@ spec:
|
|||||||
excludeRaw: true
|
excludeRaw: true
|
||||||
|
|
||||||
vaultAuthRef: llama
|
vaultAuthRef: llama
|
||||||
---
|
|
||||||
apiVersion: secrets.hashicorp.com/v1beta1
|
|
||||||
kind: VaultStaticSecret
|
|
||||||
metadata:
|
|
||||||
name: llama-openrouter
|
|
||||||
namespace: llama
|
|
||||||
spec:
|
|
||||||
type: kv-v2
|
|
||||||
|
|
||||||
mount: secret
|
|
||||||
path: openrouter
|
|
||||||
|
|
||||||
destination:
|
|
||||||
create: true
|
|
||||||
name: llama-openrouter
|
|
||||||
type: Opaque
|
|
||||||
transformation:
|
|
||||||
excludeRaw: true
|
|
||||||
templates:
|
|
||||||
OPENROUTER_API_KEY:
|
|
||||||
text: '{{ get .Secrets "API_KEY" }}'
|
|
||||||
|
|
||||||
vaultAuthRef: llama
|
|
||||||
|
|||||||
Reference in New Issue
Block a user