2 Commits

Author SHA1 Message Date
08e13bd55a Update renovate/renovate Docker tag to v41.43.7 2025-07-29 00:02:15 +00:00
257cca891a add gemma3 model 2025-07-29 01:57:40 +02:00
6 changed files with 12 additions and 79 deletions

View File

@@ -18,7 +18,7 @@ spec:
chart: chart:
spec: spec:
chart: immich chart: immich
version: 0.7.1 version: 0.7.0
sourceRef: sourceRef:
kind: HelmRepository kind: HelmRepository
name: secustor name: secustor
@@ -58,7 +58,6 @@ spec:
annotations: annotations:
cert-manager.io/cluster-issuer: letsencrypt cert-manager.io/cluster-issuer: letsencrypt
nginx.org/client-max-body-size: "0" nginx.org/client-max-body-size: "0"
nginx.org/websocket-services: immich-server
hosts: hosts:
- host: immich.lumpiasty.xyz - host: immich.lumpiasty.xyz
paths: paths:

View File

@@ -9,5 +9,5 @@ spec:
- ReadWriteOnce - ReadWriteOnce
resources: resources:
requests: requests:
storage: 150Gi storage: 50Gi
storageClassName: mayastor-single-hdd storageClassName: mayastor-single-hdd

View File

@@ -2,11 +2,11 @@
apiVersion: source.toolkit.fluxcd.io/v1 apiVersion: source.toolkit.fluxcd.io/v1
kind: HelmRepository kind: HelmRepository
metadata: metadata:
name: dynomite567-charts name: bat-librechat
namespace: librechat namespace: librechat
spec: spec:
interval: 24h interval: 24h
url: https://dynomite567.github.io/helm-charts/ url: https://charts.blue-atlas.de
--- ---
apiVersion: helm.toolkit.fluxcd.io/v2 apiVersion: helm.toolkit.fluxcd.io/v2
kind: HelmRelease kind: HelmRelease
@@ -18,10 +18,10 @@ spec:
chart: chart:
spec: spec:
chart: librechat chart: librechat
version: 1.8.9 version: 1.8.10
sourceRef: sourceRef:
kind: HelmRepository kind: HelmRepository
name: dynomite567-charts name: bat-librechat
values: values:
global: global:
librechat: librechat:
@@ -61,7 +61,7 @@ spec:
modelDisplayLabel: "Ollama" modelDisplayLabel: "Ollama"
- name: "Llama.cpp" - name: "Llama.cpp"
apiKey: "llama" apiKey: "llama"
baseURL: "http://llama.llama.svc.cluster.local:11434/v1" baseURL: "http://llama.llama.svc.cluster.local:11434/v1/chat/completions"
models: models:
default: [ default: [
"DeepSeek-R1-0528-Qwen3-8B-GGUF", "DeepSeek-R1-0528-Qwen3-8B-GGUF",
@@ -69,10 +69,7 @@ spec:
"Qwen3-8B-GGUF-no-thinking", "Qwen3-8B-GGUF-no-thinking",
"gemma3n-e4b", "gemma3n-e4b",
"gemma3-12b", "gemma3-12b",
"gemma3-12b-q2", "gemma3-12b-novision"
"gemma3-12b-novision",
"gemma3-4b",
"gemma3-4b-novision"
] ]
titleConvo: true titleConvo: true
titleModel: "current_model" titleModel: "current_model"
@@ -90,9 +87,6 @@ spec:
className: nginx className: nginx
annotations: annotations:
cert-manager.io/cluster-issuer: letsencrypt cert-manager.io/cluster-issuer: letsencrypt
nginx.org/client-max-body-size: "0"
nginx.ingress.kubernetes.io/proxy-buffering: "false"
nginx.org/proxy-read-timeout: 30m
hosts: hosts:
- host: librechat.lumpiasty.xyz - host: librechat.lumpiasty.xyz
paths: paths:

View File

@@ -1,33 +1,24 @@
healthCheckTimeout: 600
models: models:
"DeepSeek-R1-0528-Qwen3-8B-GGUF": "DeepSeek-R1-0528-Qwen3-8B-GGUF":
ttl: 600
cmd: | cmd: |
/app/llama-server /app/llama-server
-hf unsloth/DeepSeek-R1-0528-Qwen3-8B-GGUF:Q4_K_M -hf unsloth/DeepSeek-R1-0528-Qwen3-8B-GGUF:Q4_K_M
-ngl 37 -c 16384 -ngl 37 -c 16384
--no-warmup
--port ${PORT} --port ${PORT}
"Qwen3-8B-GGUF": "Qwen3-8B-GGUF":
ttl: 600
cmd: | cmd: |
/app/llama-server /app/llama-server
-hf unsloth/Qwen3-8B-GGUF:Q4_K_M -hf unsloth/Qwen3-8B-GGUF:Q4_K_M
-ngl 37 -c 16384 -ngl 37 -c 16384
--no-warmup
--port ${PORT} --port ${PORT}
"Qwen3-8B-GGUF-no-thinking": "Qwen3-8B-GGUF-no-thinking":
ttl: 600
cmd: | cmd: |
/app/llama-server /app/llama-server
-hf unsloth/Qwen3-8B-GGUF:Q4_K_M -hf unsloth/Qwen3-8B-GGUF:Q4_K_M
-ngl 37 -c 16384 -ngl 37 -c 16384
--jinja --chat-template-file /config/qwen_nothink_chat_template.jinja --jinja --chat-template-file /config/qwen_nothink_chat_template.jinja
--no-warmup
--port ${PORT} --port ${PORT}
"gemma3n-e3b": "gemma3n-e3b":
ttl: 600
cmd: | cmd: |
/app/llama-server /app/llama-server
-hf unsloth/gemma-3n-E4B-it-GGUF:UD-Q4_K_XL -hf unsloth/gemma-3n-E4B-it-GGUF:UD-Q4_K_XL
@@ -40,13 +31,11 @@ models:
--min-p 0.00 --min-p 0.00
--top-k 64 --top-k 64
--top-p 0.95 --top-p 0.95
--no-warmup
--port ${PORT} --port ${PORT}
"gemma3-12b": "gemma3-12b":
ttl: 600
cmd: | cmd: |
/app/llama-server /app/llama-server
-hf unsloth/gemma-3-12b-it-GGUF:Q4_K_M -hf unsloth/gemma-3-12b-it-GGUF:Q3_K_M
--ctx-size 16384 --ctx-size 16384
--n-gpu-layers 99 --n-gpu-layers 99
--prio 2 --prio 2
@@ -55,13 +44,11 @@ models:
--min-p 0.00 --min-p 0.00
--top-k 64 --top-k 64
--top-p 0.95 --top-p 0.95
--no-warmup
--port ${PORT} --port ${PORT}
"gemma3-12b-novision": "gemma3-12b-novision":
ttl: 600
cmd: | cmd: |
/app/llama-server /app/llama-server
-hf unsloth/gemma-3-12b-it-GGUF:Q4_K_M -hf unsloth/gemma-3-12b-it-GGUF:Q3_K_M
--ctx-size 16384 --ctx-size 16384
--n-gpu-layers 99 --n-gpu-layers 99
--prio 2 --prio 2
@@ -71,51 +58,4 @@ models:
--top-k 64 --top-k 64
--top-p 0.95 --top-p 0.95
--no-mmproj --no-mmproj
--no-warmup
--port ${PORT}
"gemma3-12b-q2":
ttl: 600
cmd: |
/app/llama-server
-hf unsloth/gemma-3-12b-it-GGUF:Q2_K_L
--ctx-size 16384
--n-gpu-layers 99
--prio 2
--temp 1.0
--repeat-penalty 1.0
--min-p 0.00
--top-k 64
--top-p 0.95
--no-warmup
--port ${PORT}
"gemma3-4b":
ttl: 600
cmd: |
/app/llama-server
-hf unsloth/gemma-3-4b-it-GGUF:Q4_K_M
--ctx-size 16384
--n-gpu-layers 99
--prio 2
--temp 1.0
--repeat-penalty 1.0
--min-p 0.00
--top-k 64
--top-p 0.95
--no-warmup
--port ${PORT}
"gemma3-4b-novision":
ttl: 600
cmd: |
/app/llama-server
-hf unsloth/gemma-3-4b-it-GGUF:Q4_K_M
--ctx-size 16384
--n-gpu-layers 99
--prio 2
--temp 1.0
--repeat-penalty 1.0
--min-p 0.00
--top-k 64
--top-p 0.95
--no-mmproj
--no-warmup
--port ${PORT} --port ${PORT}

View File

@@ -18,7 +18,7 @@ spec:
chart: chart:
spec: spec:
chart: ollama chart: ollama
version: 1.25.0 version: 1.24.0
sourceRef: sourceRef:
kind: HelmRepository kind: HelmRepository
name: ollama-helm name: ollama-helm

View File

@@ -15,7 +15,7 @@ spec:
- name: renovate - name: renovate
# Update this to the latest available and then enable Renovate on # Update this to the latest available and then enable Renovate on
# the manifest # the manifest
image: renovate/renovate:41.51.0-full image: renovate/renovate:41.43.7-full
envFrom: envFrom:
- secretRef: - secretRef:
name: renovate-gitea-token name: renovate-gitea-token