1 Commits

Author SHA1 Message Date
b6b4aea71a Update renovate/renovate Docker tag to v41.37.12 2025-07-18 00:01:01 +00:00
56 changed files with 172 additions and 1065 deletions

View File

@@ -3,17 +3,7 @@ install-router:
gen-talos-config: gen-talos-config:
mkdir -p talos/generated mkdir -p talos/generated
talosctl gen config \ talosctl gen config --with-secrets secrets.yaml --config-patch @talos/patches/controlplane.patch --config-patch @talos/patches/openebs.patch --config-patch @talos/patches/openbao.patch --config-patch @talos/patches/ollama.patch --config-patch @talos/patches/frigate.patch --config-patch @talos/patches/anapistula-delrosalae.patch --output-types controlplane -o talos/generated/anapistula-delrosalae.yaml homelab https://kube-api.homelab.lumpiasty.xyz:6443
--with-secrets secrets.yaml \
--config-patch @talos/patches/controlplane.patch \
--config-patch @talos/patches/openebs.patch \
--config-patch @talos/patches/openbao.patch \
--config-patch @talos/patches/ollama.patch \
--config-patch @talos/patches/llama.patch \
--config-patch @talos/patches/frigate.patch \
--config-patch @talos/patches/anapistula-delrosalae.patch \
--output-types controlplane -o talos/generated/anapistula-delrosalae.yaml \
homelab https://kube-api.homelab.lumpiasty.xyz:6443
talosctl gen config --with-secrets secrets.yaml --config-patch @talos/patches/controlplane.patch --output-types worker -o talos/generated/worker.yaml homelab https://kube-api.homelab.lumpiasty.xyz:6443 talosctl gen config --with-secrets secrets.yaml --config-patch @talos/patches/controlplane.patch --output-types worker -o talos/generated/worker.yaml homelab https://kube-api.homelab.lumpiasty.xyz:6443
talosctl gen config --with-secrets secrets.yaml --output-types talosconfig -o talos/generated/talosconfig homelab https://kube-api.homelab.lumpiasty.xyz:6443 talosctl gen config --with-secrets secrets.yaml --output-types talosconfig -o talos/generated/talosconfig homelab https://kube-api.homelab.lumpiasty.xyz:6443
talosctl config endpoint kube-api.homelab.lumpiasty.xyz talosctl config endpoint kube-api.homelab.lumpiasty.xyz

View File

@@ -107,8 +107,8 @@ spec:
# Uparty false positive # Uparty false positive
mask: 0.739,0.725,0.856,0.76,0.862,0.659,0.746,0.614 mask: 0.739,0.725,0.856,0.76,0.862,0.659,0.746,0.614
# ffmpeg: ffmpeg:
# hwaccel_args: preset-vaapi hwaccel_args: preset-vaapi
detectors: detectors:
ov_0: ov_0:
@@ -139,7 +139,7 @@ spec:
skipuninstall: true skipuninstall: true
config: config:
enabled: true enabled: true
size: 5Gi size: 1Gi
storageClass: mayastor-single-hdd storageClass: mayastor-single-hdd
skipuninstall: true skipuninstall: true
envFromSecrets: envFromSecrets:
@@ -147,9 +147,9 @@ spec:
ingress: ingress:
enabled: true enabled: true
ingressClassName: nginx-ingress
annotations: annotations:
cert-manager.io/cluster-issuer: letsencrypt cert-manager.io/cluster-issuer: letsencrypt
nginx.org/websocket-services: frigate
hosts: hosts:
- host: frigate.lumpiasty.xyz - host: frigate.lumpiasty.xyz
paths: paths:
@@ -164,16 +164,16 @@ spec:
kubernetes.io/hostname: anapistula-delrosalae kubernetes.io/hostname: anapistula-delrosalae
# GPU access # GPU access
# extraVolumes: extraVolumes:
# - name: dri - name: dri
# hostPath: hostPath:
# path: /dev/dri/renderD128 path: /dev/dri/renderD128
# type: CharDevice type: CharDevice
# extraVolumeMounts: extraVolumeMounts:
# - name: dri - name: dri
# mountPath: /dev/dri/renderD128 mountPath: /dev/dri/renderD128
# securityContext: securityContext:
# # Not ideal # Not ideal
# privileged: true privileged: true

View File

@@ -17,7 +17,7 @@ spec:
chart: chart:
spec: spec:
chart: gitea chart: gitea
version: 12.4.0 version: 12.1.1
sourceRef: sourceRef:
kind: HelmRepository kind: HelmRepository
name: gitea-charts name: gitea-charts
@@ -79,8 +79,8 @@ spec:
ssh: ssh:
annotations: annotations:
lbipam.cilium.io/sharing-key: gitea lbipam.cilium.io/sharing-key: gitea
lbipam.cilium.io/sharing-cross-namespace: nginx-ingress lbipam.cilium.io/sharing-cross-namespace: nginx-ingress-controller
lbipam.cilium.io/ips: 10.44.0.6,2001:470:61a3:400::6 lbipam.cilium.io/ips: 10.44.0.0,2001:470:61a3:400::1
type: LoadBalancer type: LoadBalancer
port: 22 port: 22
# Requirement for sharing ip with other service # Requirement for sharing ip with other service
@@ -89,7 +89,7 @@ spec:
ingress: ingress:
enabled: true enabled: true
className: nginx-ingress className: nginx
annotations: annotations:
cert-manager.io/cluster-issuer: letsencrypt cert-manager.io/cluster-issuer: letsencrypt
acme.cert-manager.io/http01-edit-in-place: "true" acme.cert-manager.io/http01-edit-in-place: "true"

View File

@@ -1,9 +0,0 @@
apiVersion: kustomize.config.k8s.io/v1beta1
kind: Kustomization
resources:
- namespace.yaml
- volume.yaml
- redis.yaml
- postgres-password.yaml
- postgres-cluster.yaml
- release.yaml

View File

@@ -1,4 +0,0 @@
apiVersion: v1
kind: Namespace
metadata:
name: immich

View File

@@ -1,32 +0,0 @@
---
apiVersion: postgresql.cnpg.io/v1
kind: Cluster
metadata:
name: immich-db
namespace: immich
spec:
imageName: ghcr.io/tensorchord/cloudnative-vectorchord:14-0.4.3
instances: 1
storage:
size: 10Gi
storageClass: mayastor-single-hdd
bootstrap:
initdb:
# Defaults of immich chart
database: immich
owner: immich
# We need to create custom role because default one does not allow to set up
# vectorchord extension
managed:
roles:
- name: immich
createdb: true
login: true
superuser: true
# We need to manually create secret
# https://github.com/cloudnative-pg/cloudnative-pg/issues/3788
passwordSecret:
name: immich-db-immich

View File

@@ -1,38 +0,0 @@
---
apiVersion: v1
kind: ServiceAccount
metadata:
name: immich-password
namespace: immich
---
apiVersion: secrets.hashicorp.com/v1beta1
kind: VaultAuth
metadata:
name: immich
namespace: immich
spec:
method: kubernetes
mount: kubernetes
kubernetes:
role: immich
serviceAccount: immich-password
---
apiVersion: secrets.hashicorp.com/v1beta1
kind: VaultStaticSecret
metadata:
name: immich-db
namespace: immich
spec:
type: kv-v2
mount: secret
path: immich-db
destination:
create: true
name: immich-db-immich
type: kubernetes.io/basic-auth
transformation:
excludeRaw: true
vaultAuthRef: immich

View File

@@ -1,29 +0,0 @@
---
apiVersion: source.toolkit.fluxcd.io/v1
kind: HelmRepository
metadata:
name: bitnami
namespace: immich
spec:
interval: 24h
type: "oci"
url: oci://registry-1.docker.io/bitnamicharts/
---
apiVersion: helm.toolkit.fluxcd.io/v2
kind: HelmRelease
metadata:
name: redis
namespace: immich
spec:
interval: 30m
chart:
spec:
chart: redis
version: 23.1.3
sourceRef:
kind: HelmRepository
name: bitnami
values:
global:
defaultStorageClass: mayastor-single-hdd
architecture: standalone

View File

@@ -1,69 +0,0 @@
---
apiVersion: source.toolkit.fluxcd.io/v1
kind: HelmRepository
metadata:
name: secustor
namespace: immich
spec:
interval: 24h
url: https://secustor.dev/helm-charts
---
apiVersion: helm.toolkit.fluxcd.io/v2
kind: HelmRelease
metadata:
name: immich
namespace: immich
spec:
interval: 30m
chart:
spec:
chart: immich
version: 1.0.2
sourceRef:
kind: HelmRepository
name: secustor
values:
common:
config:
vecotrExtension: vectorchord
postgres:
host: immich-db-rw
existingSecret:
enabled: true
secretName: immich-db-immich
usernameKey: username
passwordKey: password
redis:
host: redis-master
existingSecret:
enabled: true
secretName: redis
passwordKey: redis-password
server:
volumeMounts:
- mountPath: /usr/src/app/upload
name: uploads
volumes:
- name: uploads
persistentVolumeClaim:
claimName: library
machineLearning:
enabled: true
ingress:
enabled: true
className: nginx-ingress
annotations:
cert-manager.io/cluster-issuer: letsencrypt
nginx.ingress.kubernetes.io/proxy-body-size: "0"
hosts:
- host: immich.lumpiasty.xyz
paths:
- path: /
pathType: Prefix
tls:
- hosts:
- immich.lumpiasty.xyz
secretName: immich-ingress

View File

@@ -1,13 +0,0 @@
---
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
name: library
namespace: immich
spec:
accessModes:
- ReadWriteOnce
resources:
requests:
storage: 150Gi
storageClassName: mayastor-single-hdd

View File

@@ -4,9 +4,6 @@ resources:
- gitea - gitea
- registry - registry
- renovate - renovate
- ollama
- librechat - librechat
- frigate - frigate
- llama
- immich
- nas
- searxng

View File

@@ -2,11 +2,11 @@
apiVersion: source.toolkit.fluxcd.io/v1 apiVersion: source.toolkit.fluxcd.io/v1
kind: HelmRepository kind: HelmRepository
metadata: metadata:
name: dynomite567-charts name: bat-librechat
namespace: librechat namespace: librechat
spec: spec:
interval: 24h interval: 24h
url: https://dynomite567.github.io/helm-charts/ url: https://charts.blue-atlas.de
--- ---
apiVersion: helm.toolkit.fluxcd.io/v2 apiVersion: helm.toolkit.fluxcd.io/v2
kind: HelmRelease kind: HelmRelease
@@ -18,10 +18,10 @@ spec:
chart: chart:
spec: spec:
chart: librechat chart: librechat
version: 1.9.1 version: 1.8.10
sourceRef: sourceRef:
kind: HelmRepository kind: HelmRepository
name: dynomite567-charts name: bat-librechat
values: values:
global: global:
librechat: librechat:
@@ -39,32 +39,26 @@ spec:
endpoints: endpoints:
custom: custom:
- name: "Llama.cpp" - name: "Ollama"
apiKey: "llama" apiKey: "ollama"
baseURL: "http://llama.llama.svc.cluster.local:11434/v1" baseURL: "http://ollama.ollama.svc.cluster.local:11434/v1/chat/completions"
models: models:
default: [ default: [
"DeepSeek-R1-0528-Qwen3-8B-GGUF", "llama2",
"Qwen3-8B-GGUF", "mistral",
"Qwen3-8B-GGUF-no-thinking", "codellama",
"gemma3n-e4b", "dolphin-mixtral",
"gemma3-12b", "mistral-openorca"
"gemma3-12b-q2",
"gemma3-12b-novision",
"gemma3-4b",
"gemma3-4b-novision",
"Qwen3-4B-Thinking-2507",
"Qwen3-4B-Thinking-2507-long-ctx",
"Qwen2.5-VL-7B-Instruct-GGUF",
"Qwen2.5-VL-32B-Instruct-GGUF-IQ1_S",
"Qwen2.5-VL-32B-Instruct-GGUF-Q2_K_L"
] ]
# fetching list of models is supported but the `name` field must start
# with `ollama` (case-insensitive), as it does in this example.
fetch: true
titleConvo: true titleConvo: true
titleModel: "gemma3-4b-novision" titleModel: "current_model"
summarize: false summarize: false
summaryModel: "gemma3-4b-novision" summaryModel: "current_model"
forcePrompt: false forcePrompt: false
modelDisplayLabel: "Llama.cpp" modelDisplayLabel: "Ollama"
imageVolume: imageVolume:
enabled: true enabled: true
size: 10G size: 10G
@@ -72,12 +66,9 @@ spec:
storageClassName: mayastor-single-hdd storageClassName: mayastor-single-hdd
ingress: ingress:
enabled: true enabled: true
className: nginx-ingress className: nginx
annotations: annotations:
cert-manager.io/cluster-issuer: letsencrypt cert-manager.io/cluster-issuer: letsencrypt
nginx.ingress.kubernetes.io/proxy-body-size: "0"
nginx.ingress.kubernetes.io/proxy-buffering: "false"
nginx.ingress.kubernetes.io/proxy-read-timeout: 30m
hosts: hosts:
- host: librechat.lumpiasty.xyz - host: librechat.lumpiasty.xyz
paths: paths:

View File

@@ -1,216 +0,0 @@
healthCheckTimeout: 600
models:
"DeepSeek-R1-0528-Qwen3-8B-GGUF":
ttl: 600
cmd: |
/app/llama-server
-hf unsloth/DeepSeek-R1-0528-Qwen3-8B-GGUF:Q4_K_M
-ngl 37 -c 16384
--no-warmup
--port ${PORT}
"Qwen3-8B-GGUF":
ttl: 600
cmd: |
/app/llama-server
-hf unsloth/Qwen3-8B-GGUF:Q4_K_M
-ngl 37 -c 16384
--no-warmup
--port ${PORT}
"Qwen3-8B-GGUF-no-thinking":
ttl: 600
cmd: |
/app/llama-server
-hf unsloth/Qwen3-8B-GGUF:Q4_K_M
-ngl 37 -c 16384
--jinja --chat-template-file /config/qwen_nothink_chat_template.jinja
--no-warmup
--port ${PORT}
"gemma3n-e4b":
ttl: 600
cmd: |
/app/llama-server
-hf unsloth/gemma-3n-E4B-it-GGUF:UD-Q4_K_XL
--ctx-size 16384
--n-gpu-layers 99
--seed 3407
--prio 2
--temp 1.0
--repeat-penalty 1.0
--min-p 0.00
--top-k 64
--top-p 0.95
--no-warmup
--port ${PORT}
"gemma3-12b":
ttl: 600
cmd: |
/app/llama-server
-hf unsloth/gemma-3-12b-it-GGUF:Q4_K_M
--ctx-size 16384
--n-gpu-layers 99
--prio 2
--temp 1.0
--repeat-penalty 1.0
--min-p 0.00
--top-k 64
--top-p 0.95
--no-warmup
--port ${PORT}
"gemma3-12b-novision":
ttl: 600
cmd: |
/app/llama-server
-hf unsloth/gemma-3-12b-it-GGUF:Q4_K_M
--ctx-size 16384
--n-gpu-layers 99
--prio 2
--temp 1.0
--repeat-penalty 1.0
--min-p 0.00
--top-k 64
--top-p 0.95
--no-mmproj
--no-warmup
--port ${PORT}
"gemma3-12b-q2":
ttl: 600
cmd: |
/app/llama-server
-hf unsloth/gemma-3-12b-it-GGUF:Q2_K_L
--ctx-size 16384
--n-gpu-layers 99
--prio 2
--temp 1.0
--repeat-penalty 1.0
--min-p 0.00
--top-k 64
--top-p 0.95
--no-warmup
--port ${PORT}
"gemma3-4b":
ttl: 600
cmd: |
/app/llama-server
-hf unsloth/gemma-3-4b-it-GGUF:Q4_K_M
--ctx-size 16384
--n-gpu-layers 99
--prio 2
--temp 1.0
--repeat-penalty 1.0
--min-p 0.00
--top-k 64
--top-p 0.95
--no-warmup
--port ${PORT}
"gemma3-4b-novision":
ttl: 600
cmd: |
/app/llama-server
-hf unsloth/gemma-3-4b-it-GGUF:Q4_K_M
--ctx-size 16384
--n-gpu-layers 99
--prio 2
--temp 1.0
--repeat-penalty 1.0
--min-p 0.00
--top-k 64
--top-p 0.95
--no-mmproj
--no-warmup
--port ${PORT}
"Qwen3-4B-Thinking-2507":
ttl: 600
cmd: |
/app/llama-server
-hf unsloth/Qwen3-4B-Thinking-2507-GGUF:Q4_K_M
-ngl 99 -c 16384 --predict 8192
--temp 0.6
--min-p 0.00
--top-p 0.95
--top-k 20
--repeat-penalty 1.0
--no-warmup
--port ${PORT}
"Qwen3-4B-Thinking-2507-long-ctx":
ttl: 600
cmd: |
/app/llama-server
-hf unsloth/Qwen3-4B-Thinking-2507-GGUF:Q4_K_M
-ngl 99 -c 262144 --predict 81920
--temp 0.6
--min-p 0.00
--top-p 0.95
--top-k 20
--repeat-penalty 1.0
--no-warmup
--flash-attn
--cache-type-k q8_0 --cache-type-v q8_0
--port ${PORT}
"Qwen3-4B-Instruct-2507":
ttl: 600
cmd: |
/app/llama-server
-hf unsloth/Qwen3-4B-Instruct-2507-GGUF:Q4_K_M
-ngl 99 -c 16384 --predict 8192
--temp 0.7
--min-p 0.00
--top-p 0.8
--top-k 20
--repeat-penalty 1.0
--no-warmup
--port ${PORT}
"Qwen3-4B-Instruct-2507-long-ctx":
ttl: 600
cmd: |
/app/llama-server
-hf unsloth/Qwen3-4B-Instruct-2507-GGUF:Q4_K_M
-ngl 99 -c 262144 --predict 81920
--temp 0.7
--min-p 0.00
--top-p 0.8
--top-k 20
--repeat-penalty 1.0
--no-warmup
--flash-attn
--cache-type-k q8_0 --cache-type-v q8_0
--port ${PORT}
"Qwen2.5-VL-32B-Instruct-GGUF-IQ1_S":
ttl: 600
cmd: |
/app/llama-server
-hf unsloth/Qwen2.5-VL-32B-Instruct-GGUF:IQ1_S
-ngl 99 -c 16384 --predict 8192
--temp 0.7
--min-p 0.00
--top-p 0.8
--top-k 20
--repeat-penalty 1.0
--no-warmup
--port ${PORT}
"Qwen2.5-VL-32B-Instruct-GGUF-Q2_K_L":
ttl: 600
cmd: |
/app/llama-server
-hf unsloth/Qwen2.5-VL-32B-Instruct-GGUF:Q2_K_L
-ngl 99 -c 16384 --predict 8192
--temp 0.7
--min-p 0.00
--top-p 0.8
--top-k 20
--repeat-penalty 1.0
--no-warmup
--port ${PORT}
"Qwen2.5-VL-7B-Instruct-GGUF":
ttl: 600
cmd: |
/app/llama-server
-hf unsloth/Qwen2.5-VL-7B-Instruct-GGUF:Q4_K_M
-ngl 37 -c 16384 --predict 8192
--temp 0.7
--min-p 0.00
--top-p 0.8
--top-k 20
--repeat-penalty 1.0
--no-warmup
--port ${PORT}

View File

@@ -1,101 +0,0 @@
{%- if not add_generation_prompt is defined %}
{%- set add_generation_prompt = false %}
{%- endif %}
{%- set ns = namespace(is_first=false, is_tool=false, is_output_first=true, system_prompt='', is_first_sp=true, is_last_user=false) %}
{%- for message in messages %}
{%- if message['role'] == 'system' %}
{%- if ns.is_first_sp %}
{%- set ns.system_prompt = ns.system_prompt + message['content'] %}
{%- set ns.is_first_sp = false %}
{%- else %}
{%- set ns.system_prompt = ns.system_prompt + '\n\n' + message['content'] %}
{%- endif %}
{%- endif %}
{%- endfor %}
{#- Adapted from https://github.com/sgl-project/sglang/blob/main/examples/chat_template/tool_chat_template_deepseekr1.jinja #}
{%- if tools is defined and tools is not none %}
{%- set tool_ns = namespace(text='You are a helpful assistant with tool calling capabilities. ' + 'When a tool call is needed, you MUST use the following format to issue the call:\n' + '<tool▁calls▁begin><tool▁call▁begin>function<tool▁sep>FUNCTION_NAME\n' + '```json\n{"param1": "value1", "param2": "value2"}\n```<tool▁call▁end><tool▁calls▁end>\n\n' + 'Make sure the JSON is valid.' + '## Tools\n\n### Function\n\nYou have the following functions available:\n\n') %}
{%- for tool in tools %}
{%- set tool_ns.text = tool_ns.text + '\n```json\n' + (tool | tojson) + '\n```\n' %}
{%- endfor %}
{%- if ns.system_prompt|length != 0 %}
{%- set ns.system_prompt = ns.system_prompt + '\n\n' + tool_ns.text %}
{%- else %}
{%- set ns.system_prompt = tool_ns.text %}
{%- endif %}
{%- endif %}
{{- bos_token }}
{{- '/no_think' + ns.system_prompt }}
{%- set last_index = (messages|length - 1) %}
{%- for message in messages %}
{%- set content = message['content'] %}
{%- if message['role'] == 'user' %}
{%- set ns.is_tool = false -%}
{%- set ns.is_first = false -%}
{%- set ns.is_last_user = true -%}
{%- if loop.index0 == last_index %}
{{- '<User>' + content }}
{%- else %}
{{- '<User>' + content + '<Assistant>'}}
{%- endif %}
{%- endif %}
{%- if message['role'] == 'assistant' %}
{%- if '</think>' in content %}
{%- set content = (content.split('</think>')|last) %}
{%- endif %}
{%- endif %}
{%- if message['role'] == 'assistant' and message['tool_calls'] is defined and message['tool_calls'] is not none %}
{%- set ns.is_last_user = false -%}
{%- if ns.is_tool %}
{{- '<tool▁outputs▁end>'}}
{%- endif %}
{%- set ns.is_first = false %}
{%- set ns.is_tool = false -%}
{%- set ns.is_output_first = true %}
{%- for tool in message['tool_calls'] %}
{%- set arguments = tool['function']['arguments'] %}
{%- if arguments is not string %}
{%- set arguments = arguments|tojson %}
{%- endif %}
{%- if not ns.is_first %}
{%- if content is none %}
{{- '<tool▁calls▁begin><tool▁call▁begin>' + tool['type'] + '<tool▁sep>' + tool['function']['name'] + '\n' + '```json' + '\n' + arguments + '\n' + '```' + '<tool▁call▁end>'}}
}
{%- else %}
{{- content + '<tool▁calls▁begin><tool▁call▁begin>' + tool['type'] + '<tool▁sep>' + tool['function']['name'] + '\n' + '```json' + '\n' + arguments + '\n' + '```' + '<tool▁call▁end>'}}
{%- endif %}
{%- set ns.is_first = true -%}
{%- else %}
{{- '\n' + '<tool▁call▁begin>' + tool['type'] + '<tool▁sep>' + tool['function']['name'] + '\n' + '```json' + '\n' + arguments + '\n' + '```' + '<tool▁call▁end>'}}
{%- endif %}
{%- endfor %}
{{- '<tool▁calls▁end><end▁of▁sentence>'}}
{%- endif %}
{%- if message['role'] == 'assistant' and (message['tool_calls'] is not defined or message['tool_calls'] is none) %}
{%- set ns.is_last_user = false -%}
{%- if ns.is_tool %}
{{- '<tool▁outputs▁end>' + content + '<end▁of▁sentence>'}}
{%- set ns.is_tool = false -%}
{%- else %}
{{- content + '<end▁of▁sentence>'}}
{%- endif %}
{%- endif %}
{%- if message['role'] == 'tool' %}
{%- set ns.is_last_user = false -%}
{%- set ns.is_tool = true -%}
{%- if ns.is_output_first %}
{{- '<tool▁outputs▁begin><tool▁output▁begin>' + content + '<tool▁output▁end>'}}
{%- set ns.is_output_first = false %}
{%- else %}
{{- '\n<tool▁output▁begin>' + content + '<tool▁output▁end>'}}
{%- endif %}
{%- endif %}
{%- endfor -%}
{%- if ns.is_tool %}
{{- '<tool▁outputs▁end>'}}
{%- endif %}
{#- if add_generation_prompt and not ns.is_last_user and not ns.is_tool #}
{%- if add_generation_prompt and not ns.is_tool %}
{{- '<Assistant>'}}
{%- endif %}

View File

@@ -1,70 +0,0 @@
---
apiVersion: apps/v1
kind: Deployment
metadata:
name: llama-swap
namespace: llama
spec:
replicas: 1
selector:
matchLabels:
app: llama-swap
template:
metadata:
labels:
app: llama-swap
spec:
containers:
- name: llama-swap
image: ghcr.io/mostlygeek/llama-swap:v166-vulkan-b6795
imagePullPolicy: IfNotPresent
command:
- /app/llama-swap
args:
- --config=/config/config.yaml
- --watch-config
ports:
- containerPort: 8080
name: http
protocol: TCP
volumeMounts:
- name: models
mountPath: /root/.cache
- mountPath: /dev/kfd
name: kfd
- mountPath: /dev/dri
name: dri
- mountPath: /config
name: config
securityContext:
privileged: true
volumes:
- name: models
persistentVolumeClaim:
claimName: llama-models
- name: kfd
hostPath:
path: /dev/kfd
type: CharDevice
- name: dri
hostPath:
path: /dev/dri
type: Directory
- name: config
configMap:
name: llama-swap
---
apiVersion: v1
kind: Service
metadata:
name: llama
namespace: llama
spec:
type: ClusterIP
ports:
- name: http
port: 11434
targetPort: 8080
protocol: TCP
selector:
app: llama-swap

View File

@@ -1,15 +0,0 @@
apiVersion: kustomize.config.k8s.io/v1beta1
kind: Kustomization
resources:
- namespace.yaml
- secret.yaml
- auth-proxy.yaml
- ingress.yaml
- pvc.yaml
- deployment.yaml
configMapGenerator:
- name: llama-swap
namespace: llama
files:
- config.yaml=configs/config.yaml
- qwen_nothink_chat_template.jinja=configs/qwen_nothink_chat_template.jinja

View File

@@ -1,13 +0,0 @@
---
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
namespace: llama
name: llama-models
spec:
accessModes:
- ReadWriteOnce
resources:
requests:
storage: 200Gi
storageClassName: mayastor-single-ssd

View File

@@ -1,14 +0,0 @@
apiVersion: v1
kind: ConfigMap
metadata:
name: nas-sshd-config
namespace: nas
data:
00-chroot.conf: |
Subsystem sftp internal-sftp
Match User nas
ChrootDirectory /config
ForceCommand internal-sftp -d /data
AllowTcpForwarding no
X11Forwarding no
PermitTunnel no

View File

@@ -1,94 +0,0 @@
apiVersion: apps/v1
kind: Deployment
metadata:
name: nas-ssh
namespace: nas
spec:
replicas: 1
selector:
matchLabels:
app: nas-ssh
template:
metadata:
labels:
app: nas-ssh
spec:
securityContext:
fsGroup: 1000
initContainers:
- name: prepare-config
image: alpine:3.20.3
imagePullPolicy: IfNotPresent
command:
- /bin/sh
- -c
- |
set -euo pipefail
chown root:root /config
chmod 755 /config
mkdir -p /config/data
chown 1000:1000 /config/data
chmod 750 /config/data
mkdir -p /config/ssh_host_keys
chown root:root /config/ssh_host_keys
chmod 700 /config/ssh_host_keys
for key in /config/ssh_host_keys/*; do
[ -f "$key" ] || continue
chown root:root "$key"
chmod 600 "$key"
done
mkdir -p /config/sshd/sshd_config.d
cp /defaults/00-chroot.conf /config/sshd/sshd_config.d/00-chroot.conf
chown root:root /config/sshd/sshd_config.d/00-chroot.conf
chmod 644 /config/sshd/sshd_config.d/00-chroot.conf
volumeMounts:
- name: data
mountPath: /config
- name: sshd-config
mountPath: /defaults/00-chroot.conf
subPath: 00-chroot.conf
readOnly: true
containers:
- name: ssh
image: lscr.io/linuxserver/openssh-server:version-10.0_p1-r9
imagePullPolicy: IfNotPresent
env:
- name: PUID
value: "1000"
- name: PGID
value: "1000"
- name: TZ
value: Etc/UTC
- name: USER_NAME
value: nas
- name: SUDO_ACCESS
value: "false"
- name: PASSWORD_ACCESS
value: "false"
- name: LOG_STDOUT
value: "true"
- name: PUBLIC_KEY
valueFrom:
secretKeyRef:
name: nas-ssh-authorized-keys
key: public_key
ports:
- containerPort: 2222
name: ssh
protocol: TCP
volumeMounts:
- name: data
mountPath: /config
resources:
requests:
cpu: 50m
memory: 128Mi
limits:
memory: 512Mi
volumes:
- name: data
persistentVolumeClaim:
claimName: nas-data
- name: sshd-config
configMap:
name: nas-sshd-config

View File

@@ -1,4 +0,0 @@
apiVersion: v1
kind: Namespace
metadata:
name: nas

View File

@@ -1,12 +0,0 @@
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
name: nas-data
namespace: nas
spec:
accessModes:
- ReadWriteOnce
resources:
requests:
storage: 500Gi
storageClassName: mayastor-single-hdd

View File

@@ -1,9 +0,0 @@
apiVersion: v1
kind: Secret
metadata:
name: nas-ssh-authorized-keys
namespace: nas
type: Opaque
stringData:
public_key: |
ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCresbDFZijI+rZMgd3LdciPjpb4x4S5B7y0U+EoYPaz6hILT72fyz3QdcgKJJv8JUJI6g0811/yFRuOzCXgWaA922c/S/t6HMUrorh7mPVQMTN2dc/SVBvMa7S2M9NYBj6z1X2LRHs+g1JTMCtL202PIjes/E9qu0as0Vx6n/6HHNmtmA9LrpiAmurbeKXDmrYe2yWg/FA6cX5d86SJb21Dj8WqdCd3Hz0Pi6FzMKXhpWvs5Hfei1htsjsRzCxkpSTjlgFEFVfmHIXPfB06Sa6aCnkxAFnE7N+xNa9RIWeZmOXdA74LsfSKQ9eAXSrsC/IRxo2ce8cBzXJy+Itxw24fUqGYXBiCgx8i3ZA9IdwI1u71xYo9lyNjav5VykzKnAHRAYnDm9UsCf8k04reBevcLdtxL11vPCtind3xn76Nhy2b45dcp/MdYFANGsCcXJOMb6Aisb03HPGhs/aU3tCAQbTVe195mL9FWhGqIK2wBmF1SKW+4ssX2bIU6YaCYc= cardno:23_671_999

View File

@@ -1,15 +0,0 @@
apiVersion: v1
kind: Service
metadata:
name: nas-ssh
namespace: nas
spec:
type: LoadBalancer
externalTrafficPolicy: Cluster
ports:
- name: ssh
port: 22
targetPort: 2222
protocol: TCP
selector:
app: nas-ssh

View File

@@ -2,21 +2,21 @@
apiVersion: apps/v1 apiVersion: apps/v1
kind: Deployment kind: Deployment
metadata: metadata:
name: llama-proxy name: ollama-proxy
namespace: llama namespace: ollama
spec: spec:
replicas: 1 replicas: 1
selector: selector:
matchLabels: matchLabels:
app.kubernetes.io/name: llama-proxy app.kubernetes.io/name: ollama-proxy
template: template:
metadata: metadata:
labels: labels:
app.kubernetes.io/name: llama-proxy app.kubernetes.io/name: ollama-proxy
spec: spec:
containers: containers:
- name: caddy - name: caddy
image: caddy:2.10.2-alpine image: caddy:2.10.0-alpine
imagePullPolicy: IfNotPresent imagePullPolicy: IfNotPresent
volumeMounts: volumeMounts:
- mountPath: /etc/caddy - mountPath: /etc/caddy
@@ -25,21 +25,21 @@ spec:
- name: API_KEY - name: API_KEY
valueFrom: valueFrom:
secretKeyRef: secretKeyRef:
name: llama-api-key name: ollama-api-key
key: API_KEY key: API_KEY
volumes: volumes:
- name: proxy-config - name: proxy-config
configMap: configMap:
name: llama-proxy-config name: ollama-proxy-config
--- ---
apiVersion: v1 apiVersion: v1
kind: ConfigMap kind: ConfigMap
metadata: metadata:
namespace: llama namespace: ollama
name: llama-proxy-config name: ollama-proxy-config
data: data:
Caddyfile: | Caddyfile: |
http://llama.lumpiasty.xyz { http://ollama.lumpiasty.xyz {
@requireAuth { @requireAuth {
not header Authorization "Bearer {env.API_KEY}" not header Authorization "Bearer {env.API_KEY}"
@@ -47,7 +47,7 @@ data:
respond @requireAuth "Unauthorized" 401 respond @requireAuth "Unauthorized" 401
reverse_proxy llama:11434 { reverse_proxy ollama:11434 {
flush_interval -1 flush_interval -1
} }
} }
@@ -55,12 +55,12 @@ data:
apiVersion: v1 apiVersion: v1
kind: Service kind: Service
metadata: metadata:
namespace: llama namespace: ollama
name: llama-proxy name: ollama-proxy
spec: spec:
type: ClusterIP type: ClusterIP
selector: selector:
app.kubernetes.io/name: llama-proxy app.kubernetes.io/name: ollama-proxy
ports: ports:
- name: http - name: http
port: 80 port: 80

View File

@@ -2,27 +2,27 @@
apiVersion: networking.k8s.io/v1 apiVersion: networking.k8s.io/v1
kind: Ingress kind: Ingress
metadata: metadata:
namespace: llama namespace: ollama
name: llama name: ollama
annotations: annotations:
cert-manager.io/cluster-issuer: letsencrypt cert-manager.io/cluster-issuer: letsencrypt
acme.cert-manager.io/http01-edit-in-place: "true" acme.cert-manager.io/http01-edit-in-place: "true"
nginx.ingress.kubernetes.io/proxy-buffering: "false" nginx.ingress.kubernetes.io/proxy-buffering: "false"
nginx.ingress.kubernetes.io/proxy-read-timeout: 30m nginx.org/proxy-read-timeout: 30m
spec: spec:
ingressClassName: nginx-ingress ingressClassName: nginx
rules: rules:
- host: llama.lumpiasty.xyz - host: ollama.lumpiasty.xyz
http: http:
paths: paths:
- backend: - backend:
service: service:
name: llama-proxy name: ollama-proxy
port: port:
number: 80 number: 80
path: / path: /
pathType: Prefix pathType: Prefix
tls: tls:
- hosts: - hosts:
- llama.lumpiasty.xyz - ollama.lumpiasty.xyz
secretName: llama-ingress secretName: ollama-ingress

View File

@@ -2,8 +2,7 @@ apiVersion: kustomize.config.k8s.io/v1beta1
kind: Kustomization kind: Kustomization
resources: resources:
- namespace.yaml - namespace.yaml
- release.yaml
- secret.yaml - secret.yaml
- configmap.yaml - auth-proxy.yaml
- pvc.yaml - ingress.yaml
- deployment.yaml
- service.yaml

View File

@@ -2,4 +2,4 @@
apiVersion: v1 apiVersion: v1
kind: Namespace kind: Namespace
metadata: metadata:
name: llama name: ollama

60
apps/ollama/release.yaml Normal file
View File

@@ -0,0 +1,60 @@
---
apiVersion: source.toolkit.fluxcd.io/v1
kind: HelmRepository
metadata:
name: ollama-helm
namespace: ollama
spec:
interval: 24h
url: https://otwld.github.io/ollama-helm/
---
apiVersion: helm.toolkit.fluxcd.io/v2
kind: HelmRelease
metadata:
name: ollama
namespace: ollama
spec:
interval: 30m
chart:
spec:
chart: ollama
version: 1.23.0
sourceRef:
kind: HelmRepository
name: ollama-helm
namespace: ollama
interval: 12h
values:
ollama:
gpu:
enabled: false
persistentVolume:
enabled: true
storageClass: mayastor-single-hdd
size: 200Gi
# GPU support
# Rewrite of options in
# https://hub.docker.com/r/grinco/ollama-amd-apu
image:
repository: grinco/ollama-amd-apu
tag: vulkan
securityContext:
# Not ideal
privileged: true
capabilities:
add:
- PERFMON
volumeMounts:
- name: kfd
mountPath: /dev/kfd
- name: dri
mountPath: /dev/dri
volumes:
- name: kfd
hostPath:
path: /dev/kfd
type: CharDevice
- name: dri
hostPath:
path: /dev/dri
type: Directory

View File

@@ -2,26 +2,26 @@
apiVersion: v1 apiVersion: v1
kind: ServiceAccount kind: ServiceAccount
metadata: metadata:
name: llama-proxy name: ollama-proxy
namespace: llama namespace: ollama
--- ---
apiVersion: secrets.hashicorp.com/v1beta1 apiVersion: secrets.hashicorp.com/v1beta1
kind: VaultAuth kind: VaultAuth
metadata: metadata:
name: llama name: ollama
namespace: llama namespace: ollama
spec: spec:
method: kubernetes method: kubernetes
mount: kubernetes mount: kubernetes
kubernetes: kubernetes:
role: llama-proxy role: ollama-proxy
serviceAccount: llama-proxy serviceAccount: ollama-proxy
--- ---
apiVersion: secrets.hashicorp.com/v1beta1 apiVersion: secrets.hashicorp.com/v1beta1
kind: VaultStaticSecret kind: VaultStaticSecret
metadata: metadata:
name: llama-api-key name: ollama-api-key
namespace: llama namespace: ollama
spec: spec:
type: kv-v2 type: kv-v2
@@ -30,9 +30,9 @@ spec:
destination: destination:
create: true create: true
name: llama-api-key name: ollama-api-key
type: Opaque type: Opaque
transformation: transformation:
excludeRaw: true excludeRaw: true
vaultAuthRef: llama vaultAuthRef: ollama

View File

@@ -6,9 +6,9 @@ metadata:
name: registry name: registry
annotations: annotations:
cert-manager.io/cluster-issuer: letsencrypt cert-manager.io/cluster-issuer: letsencrypt
nginx.ingress.kubernetes.io/proxy-body-size: "0" nginx.org/client-max-body-size: "0"
spec: spec:
ingressClassName: nginx-ingress ingressClassName: nginx
rules: rules:
- host: registry.lumpiasty.xyz - host: registry.lumpiasty.xyz
http: http:

View File

@@ -15,7 +15,7 @@ spec:
- name: renovate - name: renovate
# Update this to the latest available and then enable Renovate on # Update this to the latest available and then enable Renovate on
# the manifest # the manifest
image: renovate/renovate:41.169.1-full image: renovate/renovate:41.37.12-full
envFrom: envFrom:
- secretRef: - secretRef:
name: renovate-gitea-token name: renovate-gitea-token

View File

@@ -1 +0,0 @@
use_default_settings: true

View File

@@ -1,42 +0,0 @@
---
apiVersion: apps/v1
kind: Deployment
metadata:
name: searxng
namespace: searxng
spec:
replicas: 1
selector:
matchLabels:
app: searxng
template:
metadata:
labels:
app: searxng
spec:
containers:
- name: searxng
image: searxng/searxng:2025.8.12-6b1516d
ports:
- containerPort: 8080
env:
- name: SEARXNG_SECRET
valueFrom:
secretKeyRef:
name: searxng-secret
key: SEARXNG_SECRET
optional: false
volumeMounts:
- name: config-volume
mountPath: /etc/searxng/settings.yml
subPath: settings.yml
readOnly: true
- name: searxng-persistent-data
mountPath: /var/cache/searxng
volumes:
- name: config-volume
configMap:
name: searxng-config
- name: searxng-persistent-data
persistentVolumeClaim:
claimName: searxng-persistent-data

View File

@@ -1,25 +0,0 @@
---
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:
namespace: searxng
name: searxng
annotations:
cert-manager.io/cluster-issuer: letsencrypt
spec:
ingressClassName: nginx-ingress
rules:
- host: searxng.lumpiasty.xyz
http:
paths:
- backend:
service:
name: searxng
port:
number: 8080
path: /
pathType: Prefix
tls:
- hosts:
- searxng.lumpiasty.xyz
secretName: searxng-ingress

View File

@@ -1,13 +0,0 @@
apiVersion: kustomize.config.k8s.io/v1beta1
kind: Kustomization
resources:
- namespace.yaml
- pvc.yaml
- deployment.yaml
- service.yaml
- ingress.yaml
configMapGenerator:
- name: searxng-config
namespace: searxng
files:
- settings.yml=configs/settings.yml

View File

@@ -1,5 +0,0 @@
---
apiVersion: v1
kind: Namespace
metadata:
name: searxng

View File

@@ -1,13 +0,0 @@
---
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
namespace: searxng
name: searxng-persistent-data
spec:
accessModes:
- ReadWriteOnce
resources:
requests:
storage: 1Gi
storageClassName: mayastor-single-ssd

View File

@@ -1,14 +0,0 @@
---
apiVersion: v1
kind: Service
metadata:
name: searxng
namespace: searxng
spec:
selector:
app: searxng
ports:
- protocol: TCP
port: 8080
targetPort: 8080
type: ClusterIP

View File

@@ -1,18 +0,0 @@
apiVersion: storage.k8s.io/v1
kind: StorageClass
metadata:
name: mayastor-single-ssd
parameters:
protocol: nvmf
# Single replica
repl: "1"
# Thin provision volumes
thin: "true"
# Generate new filesystem's uuid when cloning
cloneFsIdAsVolumeId: "true"
# Schedule this sconly on ssd
poolAffinityTopologyLabel: |
type: ssd
provisioner: io.openebs.csi-mayastor
# Allow expansion of volumes
allowVolumeExpansion: true

View File

@@ -18,14 +18,14 @@ spec:
chart: chart:
spec: spec:
chart: cert-manager-webhook-ovh chart: cert-manager-webhook-ovh
version: 0.8.0 version: 0.7.5
sourceRef: sourceRef:
kind: HelmRepository kind: HelmRepository
name: cert-manager-webhook-ovh name: cert-manager-webhook-ovh
namespace: cert-manager namespace: cert-manager
interval: 12h interval: 12h
values: values:
configVersion: 0.0.2 configVersion: 0.0.1
groupName: lumpiasty-homelab groupName: lumpiasty-homelab
certManager: certManager:
namespace: cert-manager namespace: cert-manager
@@ -38,7 +38,6 @@ spec:
acmeServerUrl: https://acme-v02.api.letsencrypt.org/directory acmeServerUrl: https://acme-v02.api.letsencrypt.org/directory
email: arek.dzski@gmail.com email: arek.dzski@gmail.com
ovhEndpointName: ovh-eu ovhEndpointName: ovh-eu
ovhAuthenticationMethod: application
ovhAuthenticationRef: ovhAuthenticationRef:
applicationKeyRef: applicationKeyRef:
name: ovh-credentials name: ovh-credentials
@@ -46,6 +45,6 @@ spec:
applicationSecretRef: applicationSecretRef:
name: ovh-credentials name: ovh-credentials
key: applicationSecret key: applicationSecret
applicationConsumerKeyRef: consumerKeyRef:
name: ovh-credentials name: ovh-credentials
key: consumerKey key: consumerKey

View File

@@ -23,7 +23,7 @@ spec:
chart: chart:
spec: spec:
chart: cert-manager chart: cert-manager
version: v1.19.1 version: v1.18.2
sourceRef: sourceRef:
kind: HelmRepository kind: HelmRepository
name: cert-manager name: cert-manager

View File

@@ -23,7 +23,7 @@ spec:
chart: chart:
spec: spec:
chart: cilium chart: cilium
version: 1.18.2 version: 1.17.5
sourceRef: sourceRef:
kind: HelmRepository kind: HelmRepository
name: cilium name: cilium

View File

@@ -23,7 +23,7 @@ spec:
chart: chart:
spec: spec:
chart: cloudnative-pg chart: cloudnative-pg
version: 0.26.0 version: 0.24.0
sourceRef: sourceRef:
kind: HelmRepository kind: HelmRepository
name: cnpg name: cnpg

View File

@@ -97,7 +97,7 @@ spec:
env: env:
- name: GOMEMLIMIT - name: GOMEMLIMIT
value: 161MiB value: 161MiB
image: registry.k8s.io/coredns/coredns:v1.13.0 image: registry.k8s.io/coredns/coredns:v1.12.2
imagePullPolicy: IfNotPresent imagePullPolicy: IfNotPresent
livenessProbe: livenessProbe:
failureThreshold: 5 failureThreshold: 5

View File

@@ -23,7 +23,7 @@ spec:
chart: chart:
spec: spec:
chart: k8up chart: k8up
version: 4.8.6 version: 4.8.4
sourceRef: sourceRef:
kind: HelmRepository kind: HelmRepository
name: k8up-io name: k8up-io

View File

@@ -2,32 +2,32 @@
apiVersion: v1 apiVersion: v1
kind: Namespace kind: Namespace
metadata: metadata:
name: nginx-ingress name: nginx-ingress-controller
--- ---
apiVersion: source.toolkit.fluxcd.io/v1 apiVersion: source.toolkit.fluxcd.io/v1
kind: HelmRepository kind: HelmRepository
metadata: metadata:
name: ingress-nginx name: nginx
namespace: nginx-ingress namespace: nginx-ingress-controller
spec: spec:
interval: 24h interval: 24h
url: https://kubernetes.github.io/ingress-nginx url: https://helm.nginx.com/stable
--- ---
apiVersion: helm.toolkit.fluxcd.io/v2 apiVersion: helm.toolkit.fluxcd.io/v2
kind: HelmRelease kind: HelmRelease
metadata: metadata:
name: nginx-ingress name: nginx-ingress
namespace: nginx-ingress namespace: nginx-ingress-controller
spec: spec:
interval: 30m interval: 30m
chart: chart:
spec: spec:
chart: ingress-nginx chart: nginx-ingress
version: 4.13.3 version: 2.1.0
sourceRef: sourceRef:
kind: HelmRepository kind: HelmRepository
name: ingress-nginx name: nginx
namespace: nginx-ingress namespace: nginx-ingress-controller
interval: 12h interval: 12h
values: values:
controller: controller:
@@ -39,11 +39,9 @@ spec:
cpu: 100m cpu: 100m
memory: 128Mi memory: 128Mi
ingressClass: "nginx-ingress" ingressClass:
ingressClassResource: create: true
name: "nginx-ingress" setAsDefaultIngress: true
enabled: true
default: false
service: service:
create: true create: true
@@ -51,11 +49,11 @@ spec:
# Requirement for sharing ip with other service # Requirement for sharing ip with other service
externalTrafficPolicy: Cluster externalTrafficPolicy: Cluster
ipFamilyPolicy: RequireDualStack ipFamilyPolicy: RequireDualStack
ipFamilies:
- IPv4
- IPv6
annotations: annotations:
# Share IP with gitea ssh so we can have the same domain for both port # Share IP with gitea ssh so we can have the same domain for both port
lbipam.cilium.io/sharing-key: gitea lbipam.cilium.io/sharing-key: gitea
lbipam.cilium.io/sharing-cross-namespace: gitea lbipam.cilium.io/sharing-cross-namespace: gitea
lbipam.cilium.io/ips: 10.44.0.6,2001:470:61a3:400::6 lbipam.cilium.io/ips: 10.44.0.0,2001:470:61a3:400::1
config:
entries:
proxy-buffering: "false"

View File

@@ -23,7 +23,7 @@ spec:
chart: chart:
spec: spec:
chart: openbao chart: openbao
version: 0.19.0 version: 0.16.1
sourceRef: sourceRef:
kind: HelmRepository kind: HelmRepository
name: openbao name: openbao

View File

@@ -23,7 +23,7 @@ spec:
chart: chart:
spec: spec:
chart: openebs chart: openebs
version: 4.3.3 version: 4.3.2
sourceRef: sourceRef:
kind: HelmRepository kind: HelmRepository
name: openebs name: openebs

View File

@@ -23,7 +23,7 @@ spec:
chart: chart:
spec: spec:
chart: vault-secrets-operator chart: vault-secrets-operator
version: 1.0.1 version: 0.10.0
sourceRef: sourceRef:
kind: HelmRepository kind: HelmRepository
name: hashicorp name: hashicorp

View File

@@ -1,11 +0,0 @@
apiVersion: "openebs.io/v1beta3"
kind: DiskPool
metadata:
name: anapistula-delrosalae-ssd
namespace: openebs
spec:
node: anapistula-delrosalae
disks: ["aio:///dev/disk/by-id/nvme-eui.000000000000000000a07501ead1ebdb"]
topology:
labelled:
type: ssd

View File

@@ -3,7 +3,7 @@ kind: Kustomization
resources: resources:
- controllers/k8up-crd-4.8.3.yaml - controllers/k8up-crd-4.8.3.yaml
- controllers/cilium.yaml - controllers/cilium.yaml
- controllers/nginx-ingress.yaml - controllers/nginx.yaml
- controllers/dns-public.yaml - controllers/dns-public.yaml
- controllers/cert-manager.yaml - controllers/cert-manager.yaml
- controllers/cert-manager-webhook-ovh.yaml - controllers/cert-manager-webhook-ovh.yaml
@@ -15,11 +15,9 @@ resources:
- controllers/mongodb-operator.yaml - controllers/mongodb-operator.yaml
- controllers/cloudnative-pg.yaml - controllers/cloudnative-pg.yaml
- diskpools/anapistula-delrosalae-hdd.yaml - diskpools/anapistula-delrosalae-hdd.yaml
- diskpools/anapistula-delrosalae-ssd.yaml
- configs/bgp-cluster-config.yaml - configs/bgp-cluster-config.yaml
- configs/loadbalancer-ippool.yaml - configs/loadbalancer-ippool.yaml
- configs/single-hdd-sc.yaml - configs/single-hdd-sc.yaml
- configs/single-ssd-sc.yaml
- configs/mayastor-snapshotclass.yaml - configs/mayastor-snapshotclass.yaml
- configs/openbao-cert.yaml - configs/openbao-cert.yaml
- configs/ovh-cert-manager-secret.yaml - configs/ovh-cert-manager-secret.yaml

View File

@@ -10,14 +10,14 @@ machine:
gateway: 2001:470:61a3:100:ffff:ffff:ffff:ffff gateway: 2001:470:61a3:100:ffff:ffff:ffff:ffff
- network: 0.0.0.0/0 - network: 0.0.0.0/0
gateway: 192.168.1.1 gateway: 192.168.1.1
mtu: 1280 mtu: 1500
install: install:
diskSelector: diskSelector:
wwid: t10.ATA SSDPR-CX400-256 GUH039914 wwid: t10.ATA SSDPR-CX400-256 GUH039914
# Generated on https://factory.talos.dev/ # Generated on https://factory.talos.dev/
# amd-ucode and amdgpu # intel-ucode and amdgpu
image: factory.talos.dev/metal-installer/9c1d1b442d73f96dcd04e81463eb20000ab014062d22e1b083e1773336bc1dd5:v1.10.6 image: factory.talos.dev/installer/06deebb947b815afa53f04c450d355d3c8bc28927a387c754db1622a0a06349e:v1.9.5
extraKernelArgs: extraKernelArgs:
- cpufreq.default_governor=performance - cpufreq.default_governor=performance
sysfs: sysfs:
@@ -27,3 +27,4 @@ machine:
devices.system.cpu.cpu6.cpufreq.scaling_max_freq: "550000" devices.system.cpu.cpu6.cpufreq.scaling_max_freq: "550000"
devices.system.cpu.cpu7.cpufreq.scaling_max_freq: "550000" devices.system.cpu.cpu7.cpufreq.scaling_max_freq: "550000"
devices.system.cpu.cpu8.cpufreq.scaling_max_freq: "550000" devices.system.cpu.cpu8.cpufreq.scaling_max_freq: "550000"

View File

@@ -1,11 +0,0 @@
# CSI driver requirement
cluster:
apiServer:
admissionControl:
- name: PodSecurity
configuration:
apiVersion: pod-security.admission.config.k8s.io/v1beta1
kind: PodSecurityConfiguration
exemptions:
namespaces:
- llama

View File

@@ -1,6 +0,0 @@
bound_service_account_names:
- immich-password
bound_service_account_namespaces:
- immich
token_policies:
- immich

View File

@@ -1,6 +0,0 @@
bound_service_account_names:
- llama-proxy
bound_service_account_namespaces:
- llama
token_policies:
- ollama

View File

@@ -1,4 +0,0 @@
path "secret/data/immich-db" {
capabilities = ["read"]
}