Homelab
This repo contains configuration and documentation for my homelab setup, which is based on Talos OS for Kubernetes cluster and MikroTik router.
Architecture
Physical setup consists of MikroTik router which connects to the internet and serves as a gateway for the cluster and other devices in the home network as shown in the diagram below.
%%{init: {"flowchart": {"ranker": "tight-tree"}}}%%
flowchart TD
subgraph internet[Internet]
ipv4[IPv4 Internet]
ipv6[IPv6 Internet]
he_tunnel[Hurricane Electric IPv6 Tunnel Broker]
isp[ISP]
end
subgraph home[Home network]
router[MikroTik Router]
cluster[Talos cluster]
lan[LAN]
mgmt[Management network]
cam[Camera system]
router --> lan
router --> cluster
router --> mgmt
router --> cam
end
ipv4 -- "Public IPv4 address" --> isp
ipv6 -- "Routed /48 IPv6 prefix" --> he_tunnel -- "6in4 Tunnel" --> isp
isp --> router
Devices are separated into VLANs and subnets for isolation and firewalling between devices and services. Whole internal network is configured to eliminate NAT where unnecessary. Pods on the Kubernetes cluster communicate with the router using native IP routing, there is no encapsulation, overlay network nor NAT on the nodes. Router knows where to direct packets destined for the pods because the cluster announces its IP prefixes to the router using BGP. Router also performs NAT for IPv4 traffic from the cluster to and from the internet, while IPv6 traffic is routed directly to the internet without NAT. High level logical routing diagram is shown below.
flowchart TD
isp[ISP] --- gpon
subgraph device[MikroTik CRS418-8P-8G-2s+]
direction TB
gpon[SFP GPON ONU]
pppoe[PPPoE client]
he_tunnel[HE Tunnel]
router[Router]@{ shape: cyl }
dockers["""
Dockers Containers (bridge)
2001:470:61a3:500::/64
172.17.0.0/16
"""]@{ shape: cloud }
tailscale["Tailscale Container"]
lan["""
LAN (vlan2)
2001:470:61a3::/64
192.168.0.0/24
"""]@{ shape: cloud }
mgmt["""
Management network (vlan1)
192.168.255.0/24
"""]@{ shape: cloud }
cam["""
Camera system (vlan3)
192.168.3.0/24
"""]@{ shape: cloud }
cluster["""
Kubernetes cluster (vlan4)
2001:470:61a3:100::/64
192.168.1.0/24
"""]@{ shape: cloud }
gpon --- pppoe -- """
139.28.40.212
Default IPv4 gateway
""" --- router
pppoe --- he_tunnel -- """
2001:470:61a3:: incoming
Default IPv6 gateway
""" --- router
router -- """
2001:470:61a3:500:ffff:ffff:ffff:ffff
172.17.0.1/16
""" --- dockers --- tailscale
router -- """
2001:470:61a3:0:ffff:ffff:ffff:ffff
192.168.0.1
"""--- lan
router -- """
192.168.255.10
"""--- mgmt
router -- "192.168.3.1" --- cam
router -- """
2001:470:61a3:100::1
192.168.1.1
""" --- cluster
end
subgraph k8s[K8s cluster]
direction TB
pod_network["""
Pod networks
2001:470:61a3:200::/104
10.42.0.0/16
(Dynamically allocated /120 IPv6 and /24 IPv4 prefixes per node)
"""]@{ shape: cloud }
service_network["""
Service network
2001:470:61a3:300::/112
10.43.0.0/16
(Advertises vIP addresses via BGP from nodes hosting endpoints)
"""]@{ shape: cloud }
load_balancer["""
Load balancer network
2001:470:61a3:400::/112
10.44.0.0/16
(Advertises vIP addresses via BGP from nodes hosting endpoints)
"""]@{ shape: cloud }
end
cluster -- "Routes exported via BGP" ----- k8s
Currently the k8s cluster consists of single node (hostname anapistula-delrosalae), which is a PC with Ryzen 5 3600, 64GB RAM, RX 580 8GB (for accelerating LLMs), 1TB NVMe SSD, 2TB and 3TB HDDs and serves both as control plane and worker node.
Software stack
The cluster itself is based on Talos Linux (which is also a Kubernetes distribution) and uses Cilium as CNI, IPAM, kube-proxy replacement, Load Balancer, and BGP control plane. Persistent volumes are managed by OpenEBS LVM LocalPV. Applications are deployed using GitOps (this repo) and reconciled on cluster using Flux. Git repository is hosted on Gitea running on a cluster itself. Secets are kept in OpenBao (HashiCorp Vault fork) running on a cluster and synced to cluster objects using Vault Secrets Operator. Deployments are kept up to date using self hosted Renovate bot updating manifests in the Git repository. There is a Woodpecker instance watching repositories on Gitea and scheduling jobs on cluster. Incoming HTTP traffic is routed to cluster using Nginx Ingress Controller and certificates are issued by cert-manager with Let's Encrypt ACME issuer with cert-manager-webhook-ovh resolving DNS-01 challanges. Cluster also runs CloudNativePG operator for managing PostgreSQL databases. Router is running Mikrotik RouterOS and its configuration is managed via Ansible playbook in this repo. High level core cluster software architecture is shown on the diagram below.
Talos Linux is an immutable Linux distribution purpose-built for running Kubernetes. The OS is distributed as an OCI (Docker) image and does not contain any package manager, shell, SSH, or any other tools for managing the system. Instead, all operations are performed using API, which can be accessed using
talosctlCLI tool.
flowchart TD
router[MikroTik Router]
router -- "Routes HTTP traffic" --> nginx
cilium -- "Announces routes via BGP" --> router
subgraph cluster[K8s cluster]
direction TB
flux[Flux CD] -- "Reconciles manifests" --> kubeapi[Kube API Server]
flux -- "Fetches Git repo" --> gitea[Gitea]
kubeapi -- "Configs, Services, Pods" --> cilium[Cilium]
cilium -- "Routing" --> services[Services] -- "Endpoints" --> pods[Pods]
cilium -- "Configures routing, interfaces, IPAM" --> pods[Pods]
kubeapi -- "Ingress rules" --> nginx[NGINX Ingress Controller] -- "Routes HTTP traffic" ---> pods
kubeapi -- "Certificate requests" --> cert_manager[cert-manager] -- "Provides certificates" --> nginx
cert_manager -- "ACME DNS-01 challanges" --> dns_webhook[cert-manager-webhook-ovh] -- "Resolves DNS challanges" --> ovh[OVH DNS]
cert_manager -- "Requests DNS-01 challanges" --> acme[Let's Encrypt ACME server] -- "Verifies domain ownership" --> ovh
kubeapi -- "Assigns pods" --> kubelet[Kubelet] -- "Manages" --> pods
kubeapi -- "PVs, LvmVols" --> openebs[OpenEBS LVM LocalPV]
openebs -- "Mounts volumes" --> pods
openebs -- "Manages" --> lv[LVM LVs]
kubeapi -- "Gets Secret refs" --> vault_operator[Vault Secrets Operator] -- "Syncs secrets" --> kubeapi
vault_operator -- "Retrieves secrets" --> vault[OpenBao] -- "Secret storage" --> lv
vault -- "Auth method" --> kubeapi
gitea -- "Receives events" --> woodpecker[Woodpecker CI] -- "Schedules jobs" --> kubeapi
gitea -- "Stores repositories" --> lv
gitea--> renovate[Renovate Bot] -- "Updates manifests" --> gitea
end
Reconcilation paths of each component
- Kubernetes manifests are reconciled using Flux triggerred by Woodpecker CI on push
- RouterOS configs are applied by Ansible
- Talos configs are applied using makefile
- Vault policies are applied by running
synchronize-vault.py
Software
Infrastructure
Operating systems
| Logo | Name | Description |
|---|---|---|
| Talos Linux | Kubernetes distribution and operating system for cluster nodes | |
| MikroTik RouterOS | Router operating system for MikroTik devices |
Configuration management
Networking
Storage
Development tools
| Logo | Name | Description |
|---|---|---|
| devenv | Tool for declarative managment of development environment using Nix | |
| Renovate | Bot for keeping dependencies up to date | |
| Woodpecker CI | Continous Integration system |
AI infrastructure
| Logo | Name | Address | Description |
|---|---|---|---|
| LLaMA.cpp | https://llama.lumpiasty.xyz/ | LLM inference server running local models with GPU acceleration |
Applications/Services
| Logo | Name | Address | Description |
|---|---|---|---|
| Gitea | https://gitea.lumpiasty.xyz/ | Private Git repository hosting and artifact storage (Docker, Helm charts) | |
![]() |
Open WebUI | https://openwebui.lumpiasty.xyz/ | Web UI for chatting with LLMs running on the cluster |
| iSpeak3.pl | ts3server://ispeak3.pl | Public TeamSpeak 3 voice communication server | |
| Immich | https://immich.lumpiasty.xyz/ | Self-hosted photo and video backup and streaming service | |
| Frigate | https://frigate.lumpiasty.xyz/ | NVR for camera system with AI object detection and classification |
Development
This repo leverages devenv for easy setup of a development environment. Install devenv, clone this repo and run devenv shell to make the tools and enviornment variables available in your shell. Alternatively, you can use direnv to automate enabling enviornment after entering directory in your shell. You can also install direnv extension in VSCode to automatically set up environment after opening workspace so all the fancy intellisense and extensions detect stuff correctly.
App deployment
This repo is being watched by Flux running on cluster. To change config/add new app, simply commit to this repo and wait a while for cluster to reconcile changes. You can speed up this process by "notifying" Flux using flux reconcile source git flux-system.
Flux watches 3 kustomizations in this repo:
- flux-system - cluster/flux-system directory, contains flux manifests
- infra - infra directory, contains cluster infrastructure manifests like storage classes, network policies, monitoring etc.
- apps - apps directory, contains manifests for applications deployed on cluster
Talos config changes
Talos config in this repo is stored as yaml patches under talos/patches directory. Those patches can then be compiled into full Talos config files using make gen-talos-config command. Full config can then be applied to cluster using make apply-talos-config command, which applies config to all nodes in cluster.
To compile config, you need to have secrets file, which contains certificates and keys for cluster. Those secrets are then incorporated into final config files. That is also why we can not store full config in repo.
Router config changes
Router config is stored as Ansible playbook under ansible/ directory. To apply changes to router, run ansible-playbook playbooks/routeros.yml command in ansible/ directory Before running playbook, you can check what changes will be applied to router using --check flag to ansible-playbook command, which will run playbook in "check mode" and show you the changes that would be applied without actually applying them. This is useful for verifying that your changes are correct before applying them to the router.
To run Ansible playbook, you need to have required Ansible collections installed. You can install them using ansible-galaxy collection install -r ansible/requirements.yml command. Configuring this in devenv is yet to be done, so you might need to install collections manually for now.
Secrets needed to access the router API are stored in OpenBao and loaded on demand when running playbook so you need to have access to appropriate secrets.
Kube API access
To generate kubeconfig for accessing cluster API, run make get-kubeconfig command, which will generate kubeconfig under talos/generated/kubeconfig path. Devenv automatically sets KUBECONFIG enviornment variable to point to this file, so you can start using kubectl right away.
Like above, you need secrets file to generate kubeconfig.
