Compare commits

..

1 Commits

Author SHA1 Message Date
Gitea Actions Bot
71b0eed6b1 Auto-update README with current k8s applications
All checks were successful
Terraform / Terraform (pull_request) Successful in 43s
Generated by CI/CD workflow on 2025-11-24 16:42:28

This PR updates the README.md file with the current list of applications found in the k8s/ directory structure.
2025-11-24 16:42:28 +00:00
70 changed files with 194 additions and 1695 deletions

View File

@@ -30,29 +30,21 @@ jobs:
cli_config_credentials_token: ${{ secrets.TF_API_TOKEN }} cli_config_credentials_token: ${{ secrets.TF_API_TOKEN }}
- name: Terraform Init - name: Terraform Init
env:
TF_VAR_authentik_token: ${{ secrets.AUTHENTIK_TOKEN }}
run: terraform init run: terraform init
working-directory: ./terraform/authentik working-directory: ./terraform/authentik
- name: Terraform Format - name: Terraform Format
env:
TF_VAR_authentik_token: ${{ secrets.AUTHENTIK_TOKEN }}
run: terraform fmt -check run: terraform fmt -check
continue-on-error: true continue-on-error: true
working-directory: ./terraform/authentik working-directory: ./terraform/authentik
- name: Terraform Apply - name: Terraform Apply
env:
TF_VAR_authentik_token: ${{ secrets.AUTHENTIK_TOKEN }}
run: terraform apply -var-file proxy-apps.tfvars -var-file oauth2-apps.tfvars -var-file terraform.tfvars -var-file groups.tfvars -input=false -auto-approve -parallelism=100 run: terraform apply -var-file proxy-apps.tfvars -var-file oauth2-apps.tfvars -var-file terraform.tfvars -var-file groups.tfvars -input=false -auto-approve -parallelism=100
working-directory: ./terraform/authentik working-directory: ./terraform/authentik
- name: Generate Wiki Content - name: Generate Wiki Content
if: success() if: success()
continue-on-error: true continue-on-error: true
env:
TF_VAR_authentik_token: ${{ secrets.AUTHENTIK_TOKEN }}
run: | run: |
echo "📋 Starting Wiki generation..." echo "📋 Starting Wiki generation..."
cd ./terraform/authentik cd ./terraform/authentik

View File

@@ -22,13 +22,12 @@ jobs:
- name: Install Python dependencies - name: Install Python dependencies
run: | run: |
python3 -m venv .venv pip install pyyaml
.venv/bin/pip install pyyaml
- name: Generate K8s Services Wiki - name: Generate K8s Services Wiki
run: | run: |
echo "📋 Starting K8s wiki generation..." echo "📋 Starting K8s wiki generation..."
.venv/bin/python .gitea/scripts/generate-k8s-wiki.py k8s/ Kubernetes-Services.md python3 .gitea/scripts/generate-k8s-wiki.py k8s/ Kubernetes-Services.md
if [ -f "Kubernetes-Services.md" ]; then if [ -f "Kubernetes-Services.md" ]; then
echo "✅ Wiki content generated successfully" echo "✅ Wiki content generated successfully"

View File

@@ -44,7 +44,6 @@ ArgoCD homelab project
| **jellyfin** | [![jellyfin](https://ag.hexor.cy/api/badge?name=jellyfin&revision=true)](https://ag.hexor.cy/applications/argocd/jellyfin) | | **jellyfin** | [![jellyfin](https://ag.hexor.cy/api/badge?name=jellyfin&revision=true)](https://ag.hexor.cy/applications/argocd/jellyfin) |
| **k8s-secrets** | [![k8s-secrets](https://ag.hexor.cy/api/badge?name=k8s-secrets&revision=true)](https://ag.hexor.cy/applications/argocd/k8s-secrets) | | **k8s-secrets** | [![k8s-secrets](https://ag.hexor.cy/api/badge?name=k8s-secrets&revision=true)](https://ag.hexor.cy/applications/argocd/k8s-secrets) |
| **khm** | [![khm](https://ag.hexor.cy/api/badge?name=khm&revision=true)](https://ag.hexor.cy/applications/argocd/khm) | | **khm** | [![khm](https://ag.hexor.cy/api/badge?name=khm&revision=true)](https://ag.hexor.cy/applications/argocd/khm) |
| **n8n** | [![n8n](https://ag.hexor.cy/api/badge?name=n8n&revision=true)](https://ag.hexor.cy/applications/argocd/n8n) |
| **ollama** | [![ollama](https://ag.hexor.cy/api/badge?name=ollama&revision=true)](https://ag.hexor.cy/applications/argocd/ollama) | | **ollama** | [![ollama](https://ag.hexor.cy/api/badge?name=ollama&revision=true)](https://ag.hexor.cy/applications/argocd/ollama) |
| **paperless** | [![paperless](https://ag.hexor.cy/api/badge?name=paperless&revision=true)](https://ag.hexor.cy/applications/argocd/paperless) | | **paperless** | [![paperless](https://ag.hexor.cy/api/badge?name=paperless&revision=true)](https://ag.hexor.cy/applications/argocd/paperless) |
| **pasarguard** | [![pasarguard](https://ag.hexor.cy/api/badge?name=pasarguard&revision=true)](https://ag.hexor.cy/applications/argocd/pasarguard) | | **pasarguard** | [![pasarguard](https://ag.hexor.cy/api/badge?name=pasarguard&revision=true)](https://ag.hexor.cy/applications/argocd/pasarguard) |
@@ -57,7 +56,6 @@ ArgoCD homelab project
| **tg-bots** | [![tg-bots](https://ag.hexor.cy/api/badge?name=tg-bots&revision=true)](https://ag.hexor.cy/applications/argocd/tg-bots) | | **tg-bots** | [![tg-bots](https://ag.hexor.cy/api/badge?name=tg-bots&revision=true)](https://ag.hexor.cy/applications/argocd/tg-bots) |
| **vaultwarden** | [![vaultwarden](https://ag.hexor.cy/api/badge?name=vaultwarden&revision=true)](https://ag.hexor.cy/applications/argocd/vaultwarden) | | **vaultwarden** | [![vaultwarden](https://ag.hexor.cy/api/badge?name=vaultwarden&revision=true)](https://ag.hexor.cy/applications/argocd/vaultwarden) |
| **vpn** | [![vpn](https://ag.hexor.cy/api/badge?name=vpn&revision=true)](https://ag.hexor.cy/applications/argocd/vpn) | | **vpn** | [![vpn](https://ag.hexor.cy/api/badge?name=vpn&revision=true)](https://ag.hexor.cy/applications/argocd/vpn) |
| **xandikos** | [![xandikos](https://ag.hexor.cy/api/badge?name=xandikos&revision=true)](https://ag.hexor.cy/applications/argocd/xandikos) |
</td> </td>
</tr> </tr>

View File

@@ -77,8 +77,8 @@ spec:
labels: labels:
app: gitea-runner app: gitea-runner
spec: spec:
#nodeSelector: nodeSelector:
# kubernetes.io/hostname: home.homenet kubernetes.io/hostname: home.homenet
volumes: volumes:
- name: docker-sock - name: docker-sock
hostPath: hostPath:
@@ -90,30 +90,27 @@ spec:
affinity: affinity:
nodeAffinity: nodeAffinity:
preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution:
- weight: 1
preference:
matchExpressions:
- key: kubernetes.io/hostname
operator: In
values:
- home.homenet
- weight: 2
preference:
matchExpressions:
- key: kubernetes.io/hostname
operator: In
values:
- master.tail2fe2d.ts.net
- weight: 3 - weight: 3
preference: preference:
matchExpressions: matchExpressions:
- key: kubernetes.io/hostname - key: kubernetes.io/hostname
operator: In operator: In
values: values:
- it.tail2fe2d.ts.net - home.homenet
- ch.tail2fe2d.ts.net - weight: 1
- us.tail2fe2d.ts.net preference:
matchExpressions:
- key: kubernetes.io/hostname
operator: In
values:
- master.tail2fe2d.ts.net
- weight: 2
preference:
matchExpressions:
- key: kubernetes.io/hostname
operator: In
values:
- nas.homenet
requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution:
nodeSelectorTerms: nodeSelectorTerms:
- matchExpressions: - matchExpressions:
@@ -121,9 +118,7 @@ spec:
operator: In operator: In
values: values:
- home.homenet - home.homenet
- it.tail2fe2d.ts.net - nas.homenet
- ch.tail2fe2d.ts.net
- us.tail2fe2d.ts.net
- master.tail2fe2d.ts.net - master.tail2fe2d.ts.net
containers: containers:
- name: gitea-runner - name: gitea-runner

View File

@@ -74,14 +74,19 @@ spec:
- nas.homenet - nas.homenet
volumes: volumes:
- name: upload-storage - name: upload-storage
persistentVolumeClaim: nfs:
claimName: immich-upload-pvc server: nas.homenet
path: /mnt/storage/Storage/k8s/immich/library/
readOnly: false
- name: gphoto-storage - name: gphoto-storage
persistentVolumeClaim: nfs:
claimName: immich-gphoto-pvc server: nas.homenet
path: /mnt/storage/Storage/k8s/immich/GPHOTO/
readOnly: false
- name: camera - name: camera
persistentVolumeClaim: nfs:
claimName: immich-camera-pvc server: nas.homenet
path: /mnt/storage/Storage/Syncthing-repos/PhoneCamera/
readOnly: true readOnly: true
- name: localtime - name: localtime
hostPath: hostPath:

View File

@@ -1,52 +1,79 @@
--- ---
apiVersion: v1 apiVersion: v1
kind: PersistentVolume
metadata:
name: immich-upload-pv
spec:
capacity:
storage: 500Gi
accessModes:
- ReadWriteOnce
hostPath:
path: /mnt/storage/Storage/k8s/immich/library
---
apiVersion: v1
kind: PersistentVolumeClaim kind: PersistentVolumeClaim
metadata: metadata:
name: immich-upload-pvc name: immich-upload-pvc
namespace: immich namespace: immich
spec: spec:
storageClassName: ""
accessModes: accessModes:
- ReadWriteMany - ReadWriteOnce
storageClassName: nfs-csi volumeName: immich-upload-pv
resources: resources:
requests: requests:
storage: 500Gi storage: 500Gi
--- ---
apiVersion: v1 apiVersion: v1
kind: PersistentVolume
metadata:
name: immich-gphoto-pv
spec:
capacity:
storage: 500Gi
accessModes:
- ReadWriteOnce
hostPath:
path: /mnt/storage/Storage/k8s/immich/GPHOTO
---
apiVersion: v1
kind: PersistentVolumeClaim kind: PersistentVolumeClaim
metadata: metadata:
name: immich-gphoto-pvc name: immich-gphoto-pvc
namespace: immich namespace: immich
spec: spec:
storageClassName: ""
accessModes: accessModes:
- ReadWriteMany - ReadWriteOnce
storageClassName: nfs-csi volumeName: immich-gphoto-pv
resources: resources:
requests: requests:
storage: 500Gi storage: 500Gi
--- ---
apiVersion: v1 apiVersion: v1
kind: PersistentVolume
metadata:
name: immich-db-pv
spec:
capacity:
storage: 10Gi
accessModes:
- ReadWriteOnce
hostPath:
path: /mnt/storage/Storage/k8s/immich/db
---
apiVersion: v1
kind: PersistentVolumeClaim kind: PersistentVolumeClaim
metadata: metadata:
name: immich-db-pvc name: immich-db-pvc
namespace: immich namespace: immich
spec: spec:
storageClassName: ""
accessModes: accessModes:
- ReadWriteMany - ReadWriteOnce
storageClassName: nfs-csi volumeName: immich-db-pv
resources: resources:
requests: requests:
storage: 10Gi storage: 10Gi
---
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
name: immich-camera-pvc
namespace: immich
spec:
accessModes:
- ReadOnlyMany
storageClassName: nfs-csi
resources:
requests:
storage: 100Gi

View File

@@ -1,5 +1,5 @@
image: image:
tag: 10.11.4 tag: 10.10.7
resources: resources:
requests: requests:
memory: "2Gi" memory: "2Gi"
@@ -36,40 +36,8 @@ ingress:
paths: paths:
- path: / - path: /
pathType: Prefix pathType: Prefix
- host: us.hexor.cy
paths:
- path: /
pathType: Prefix
- host: ch.hexor.cy
paths:
- path: /
pathType: Prefix
- host: jp.hexor.cy
paths:
- path: /
pathType: Prefix
- host: spb.hexor.cy
paths:
- path: /
pathType: Prefix
- host: cy.hexor.cy
paths:
- path: /
pathType: Prefix
- host: am.hexor.cy
paths:
- path: /
pathType: Prefix
- host: de.hexor.cy
paths:
- path: /
pathType: Prefix
- host: it.hexor.cy
paths:
- path: /
pathType: Prefix
tls: tls:
- secretName: jellyfin-tls - secretName: jellyfin-tls
hosts: hosts:
- '*.hexor.cy' - 'jf.hexor.cy'

View File

@@ -19,7 +19,7 @@ spec:
kubernetes.io/os: linux kubernetes.io/os: linux
containers: containers:
- name: secret-reader - name: secret-reader
image: ultradesu/k8s-secrets:0.2.1 image: ultradesu/k8s-secrets:0.1.1
imagePullPolicy: Always imagePullPolicy: Always
args: args:
- "--secrets" - "--secrets"
@@ -28,7 +28,6 @@ spec:
- "k8s-secret" - "k8s-secret"
- "--port" - "--port"
- "3000" - "3000"
- "--webhook"
ports: ports:
- containerPort: 3000 - containerPort: 3000
name: http name: http

View File

@@ -1,21 +0,0 @@
apiVersion: argoproj.io/v1alpha1
kind: Application
metadata:
name: n8n
namespace: argocd
spec:
project: apps
destination:
namespace: n8n
server: https://kubernetes.default.svc
source:
repoURL: ssh://git@gt.hexor.cy:30022/ab/homelab.git
targetRevision: HEAD
path: k8s/apps/n8n
syncPolicy:
automated:
selfHeal: true
prune: true
syncOptions:
- CreateNamespace=true

View File

@@ -1,37 +0,0 @@
---
apiVersion: external-secrets.io/v1
kind: ExternalSecret
metadata:
name: credentials
spec:
target:
name: credentials
deletionPolicy: Delete
template:
type: Opaque
data:
postgres-password: "{{ .psql | trim }}"
N8N_ENCRYPTION_KEY: "{{ .enc_pass | trim }}"
data:
- secretKey: psql
sourceRef:
storeRef:
name: vaultwarden-login
kind: ClusterSecretStore
remoteRef:
conversionStrategy: Default
decodingStrategy: None
metadataPolicy: None
key: 2a9deb39-ef22-433e-a1be-df1555625e22
property: fields[13].value
- secretKey: enc_pass
sourceRef:
storeRef:
name: vaultwarden-login
kind: ClusterSecretStore
remoteRef:
conversionStrategy: Default
decodingStrategy: None
metadataPolicy: None
key: 18c92d73-9637-4419-8642-7f7b308460cb
property: fields[0].value

View File

@@ -1,22 +0,0 @@
apiVersion: kustomize.config.k8s.io/v1beta1
kind: Kustomization
resources:
- external-secrets.yaml
- storage.yaml
helmCharts:
- name: n8n
repo: https://community-charts.github.io/helm-charts
version: 1.16.28
releaseName: n8n
namespace: n8n
valuesFile: values-n8n.yaml
includeCRDs: true
- name: searxng
repo: https://unknowniq.github.io/helm-charts/
version: 0.1.3
releaseName: searxng
namespace: n8n
valuesFile: values-searxng.yaml
includeCRDs: true

View File

@@ -1,12 +0,0 @@
---
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
name: n8n-home
spec:
accessModes:
- ReadWriteMany
storageClassName: nfs-csi
resources:
requests:
storage: 10Gi

View File

@@ -1,53 +0,0 @@
nodeSelector:
kubernetes.io/hostname: master.tail2fe2d.ts.net
db:
type: postgresdb
main:
resources:
requests:
cpu: 100m
memory: 128Mi
limits:
cpu: 512m
memory: 512Mi
persistence:
enabled: true
existingClaim: n8n-home
mountPath: /home/node/.n8n
worker:
mode: regular
webhook:
url: https://n8n.hexor.cy
redis:
enabled: true
existingEncryptionKeySecret: credentials
externalPostgresql:
existingSecret: credentials
host: "psql.psql.svc"
username: "n8n"
database: "n8n"
ingress:
enabled: true
className: traefik
annotations:
cert-manager.io/cluster-issuer: letsencrypt
traefik.ingress.kubernetes.io/router.middlewares: kube-system-https-redirect@kubernetescrd
hosts:
- host: n8n.hexor.cy
paths:
- path: /
pathType: Prefix
tls:
- secretName: n8n-tls
hosts:
- '*.hexor.cy'

View File

@@ -1,26 +0,0 @@
config:
general:
instance_name: "HexorSearXNG"
debug: true
server:
limiter: false
public_instance: false
method: "POST"
search:
safe_search: 0
# Create limiter.toml to bypass bot detection
extraConfig: |
[botdetection.ip_lists]
pass_ip = [
'0.0.0.0/0', # Allow all IPs
'::0/0' # Allow all IPv6
]
[botdetection.ip_limit]
filter_link_local = false
link_token = false
valkey:
enabled: true
nodeSelector:
kubernetes.io/hostname: master.tail2fe2d.ts.net

View File

@@ -4,7 +4,6 @@ kind: Kustomization
resources: resources:
- app.yaml - app.yaml
- external-secrets.yaml - external-secrets.yaml
- paperless-ai.yaml
helmCharts: helmCharts:
- name: paperless-ngx - name: paperless-ngx
@@ -28,11 +27,4 @@ helmCharts:
namespace: paperless namespace: paperless
valuesFile: gotenberg-values.yaml valuesFile: gotenberg-values.yaml
includeCRDs: true includeCRDs: true
#- name: redis
# repo: oci://registry-1.docker.io/bitnamicharts/redis
# version: 24.1.0
# releaseName: redis
# namespace: paperless
# includeCRDs: true
# valuesFile: bazarr-values.yaml

View File

@@ -1,101 +0,0 @@
---
apiVersion: apps/v1
kind: Deployment
metadata:
name: paperless-ai
labels:
app: paperless-ai
spec:
replicas: 1
selector:
matchLabels:
app: paperless-ai
template:
metadata:
labels:
app: paperless-ai
spec:
nodeSelector:
kubernetes.io/hostname: nas.homenet
containers:
- name: paperless-ai
image: clusterzx/paperless-ai:latest
imagePullPolicy: Always
ports:
- containerPort: 3000
name: http
env:
- name: NODE_ENV
value: production
- name: PAPERLESS_AI_PORT
value: "3000"
resources:
requests:
memory: 512Mi
cpu: 500m
limits:
memory: 1024Mi
cpu: 2000m
#livenessProbe:
# httpGet:
# path: /
# port: 8000
# initialDelaySeconds: 30
# periodSeconds: 10
#readinessProbe:
# httpGet:
# path: /
# port: 8000
# initialDelaySeconds: 5
# periodSeconds: 5
volumeMounts:
- name: data
mountPath: /app/data
volumes:
- name: data
hostPath:
path: /mnt/storage/Storage/k8s/paperless/ai-data
type: DirectoryOrCreate
---
apiVersion: v1
kind: Service
metadata:
name: paperless-ai
namespace: paperless
labels:
app: paperless-ai
spec:
type: ClusterIP
ports:
- port: 3000
targetPort: 3000
protocol: TCP
name: http
selector:
app: paperless-ai
---
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:
name: paperless-ai-ingress
annotations:
ingressClassName: traefik
cert-manager.io/cluster-issuer: letsencrypt
traefik.ingress.kubernetes.io/router.middlewares: kube-system-https-redirect@kubernetescrd
acme.cert-manager.io/http01-edit-in-place: "true"
spec:
rules:
- host: ai-docs.hexor.cy
http:
paths:
- path: /
pathType: Prefix
backend:
service:
name: paperless-ai
port:
number: 3000
tls:
- secretName: docs-tls
hosts:
- '*.hexor.cy'

View File

@@ -1,5 +1,5 @@
image: image:
tag: 2.20.3 tag: 2.19.3
resources: resources:
requests: requests:
memory: "1Gi" memory: "1Gi"
@@ -107,8 +107,6 @@ persistence:
- path: /usr/src/paperless/consume - path: /usr/src/paperless/consume
redis: redis:
enabled: true enabled: true
image:
tag: latest
master: master:
nodeSelector: nodeSelector:
kubernetes.io/hostname: nas.homenet kubernetes.io/hostname: nas.homenet

View File

@@ -112,8 +112,47 @@ spec:
- name: scripts - name: scripts
mountPath: /scripts mountPath: /scripts
containers: containers:
- name: xray-exporter
image: alpine:3.18
imagePullPolicy: IfNotPresent
command:
- /bin/sh
- /scripts/exporter-start.sh
ports:
- name: metrics
containerPort: 9550
protocol: TCP
livenessProbe:
httpGet:
path: /scrape
port: metrics
initialDelaySeconds: 60
periodSeconds: 30
timeoutSeconds: 10
failureThreshold: 3
readinessProbe:
httpGet:
path: /scrape
port: metrics
initialDelaySeconds: 45
periodSeconds: 10
timeoutSeconds: 5
failureThreshold: 3
resources:
requests:
memory: "64Mi"
cpu: "50m"
limits:
memory: "128Mi"
cpu: "150m"
volumeMounts:
- name: shared-data
mountPath: /shared
readOnly: true
- name: scripts
mountPath: /scripts
- name: pasarguard-node - name: pasarguard-node
image: 'pasarguard/node:v0.2.1' image: 'pasarguard/node:v0.1.1'
imagePullPolicy: Always imagePullPolicy: Always
command: command:
- /bin/sh - /bin/sh
@@ -162,56 +201,16 @@ spec:
resources: resources:
requests: requests:
memory: "128Mi" memory: "128Mi"
#cpu: "500m" cpu: "100m"
limits: limits:
memory: "512Mi" memory: "512Mi"
#cpu: "1200m" cpu: "750m"
volumeMounts: volumeMounts:
- name: shared-data - name: shared-data
mountPath: /shared mountPath: /shared
readOnly: false readOnly: false
- name: scripts - name: scripts
mountPath: /scripts mountPath: /scripts
- name: xray-exporter
image: alpine:3.18
imagePullPolicy: IfNotPresent
command:
- /bin/sh
- /scripts/exporter-start.sh
ports:
- name: metrics
containerPort: 9550
protocol: TCP
livenessProbe:
httpGet:
path: /scrape
port: metrics
initialDelaySeconds: 60
periodSeconds: 30
timeoutSeconds: 10
failureThreshold: 3
readinessProbe:
httpGet:
path: /scrape
port: metrics
initialDelaySeconds: 45
periodSeconds: 10
timeoutSeconds: 5
failureThreshold: 3
resources:
requests:
memory: "64Mi"
cpu: "50m"
limits:
memory: "128Mi"
cpu: "500m"
volumeMounts:
- name: shared-data
mountPath: /shared
readOnly: true
- name: scripts
mountPath: /scripts
volumes: volumes:
- name: shared-data - name: shared-data
emptyDir: {} emptyDir: {}

View File

@@ -34,7 +34,7 @@ spec:
mountPath: /templates/subscription mountPath: /templates/subscription
containers: containers:
- name: pasarguard-web - name: pasarguard-web
image: 'pasarguard/panel:latest' image: 'pasarguard/panel:v1.7.2'
imagePullPolicy: Always imagePullPolicy: Always
envFrom: envFrom:
- secretRef: - secretRef:

View File

@@ -2,7 +2,7 @@
apiVersion: networking.k8s.io/v1 apiVersion: networking.k8s.io/v1
kind: Ingress kind: Ingress
metadata: metadata:
name: user-ui name: remnawave-tls-ingress
annotations: annotations:
ingressClassName: traefik ingressClassName: traefik
cert-manager.io/cluster-issuer: letsencrypt cert-manager.io/cluster-issuer: letsencrypt
@@ -31,7 +31,7 @@ spec:
port: port:
number: 3010 number: 3010
tls: tls:
- secretName: remnawave-user-ui-tls - secretName: remnawave-tls
hosts: hosts:
- sub.hexor.cy - sub.hexor.cy
- sub.hexor.ru - sub.hexor.ru

View File

@@ -8,5 +8,4 @@ resources:
- ./subscription-page-configmap.yaml - ./subscription-page-configmap.yaml
- ./subscription-page-deployment.yaml - ./subscription-page-deployment.yaml
- ./servicemonitor.yaml - ./servicemonitor.yaml
- ./user-ui-ingress.yaml - ./ingress.yaml
- ./panel-ingress.yaml

View File

@@ -1,37 +0,0 @@
---
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:
name: panel-ui
annotations:
ingressClassName: traefik
cert-manager.io/cluster-issuer: letsencrypt
traefik.ingress.kubernetes.io/router.middlewares: kube-system-https-redirect@kubernetescrd
acme.cert-manager.io/http01-edit-in-place: "true"
spec:
rules:
- host: rw.hexor.cy
http:
paths:
- path: /
pathType: Prefix
backend:
service:
name: remnawave
port:
number: 3000
- host: rw.hexor.ru
http:
paths:
- path: /
pathType: Prefix
backend:
service:
name: remnawave
port:
number: 3000
tls:
- secretName: remnawave-panel-tls
hosts:
- rw.hexor.cy
- rw.hexor.ru

View File

@@ -7,7 +7,7 @@ metadata:
app: remnawave-subscription-page app: remnawave-subscription-page
data: data:
APP_PORT: "3010" APP_PORT: "3010"
REMNAWAVE_PANEL_URL: "https://rw.hexor.cy" REMNAWAVE_PANEL_URL: "http://remnawave.remnawave.svc:3000"
META_TITLE: "RemnaWave Subscription" META_TITLE: "RemnaWave Subscription"
META_DESCRIPTION: "Your VPN subscription portal" META_DESCRIPTION: "Your VPN subscription portal"
META_KEYWORDS: "vpn,subscription,remnawave" META_KEYWORDS: "vpn,subscription,remnawave"

View File

@@ -28,6 +28,22 @@ spec:
- name: http - name: http
containerPort: 3010 containerPort: 3010
protocol: TCP protocol: TCP
livenessProbe:
httpGet:
path: /
port: 3010
initialDelaySeconds: 30
periodSeconds: 30
timeoutSeconds: 5
failureThreshold: 3
readinessProbe:
httpGet:
path: /
port: 3010
initialDelaySeconds: 10
periodSeconds: 10
timeoutSeconds: 3
failureThreshold: 3
resources: resources:
requests: requests:
memory: "64Mi" memory: "64Mi"

View File

@@ -16,18 +16,18 @@ helmCharts:
valuesFile: syncthing-master.yaml valuesFile: syncthing-master.yaml
includeCRDs: true includeCRDs: true
- name: syncthing
repo: https://k8s-home-lab.github.io/helm-charts
version: 4.0.0
releaseName: syncthing-khv
namespace: syncthing
valuesFile: syncthing-khv.yaml
includeCRDs: true
- name: syncthing - name: syncthing
repo: https://k8s-home-lab.github.io/helm-charts repo: https://k8s-home-lab.github.io/helm-charts
version: 4.0.0 version: 4.0.0
releaseName: syncthing-nas releaseName: syncthing-nas
namespace: syncthing namespace: syncthing
valuesFile: syncthing-nas.yaml valuesFile: syncthing-nas.yaml
includeCRDs: true includeCRDs: true
# - name: syncthing
# repo: https://k8s-home-lab.github.io/helm-charts
# version: 4.0.0
# releaseName: syncthing-khv
# namespace: syncthing
# valuesFile: syncthing-khv.yaml
# includeCRDs: true

View File

@@ -1,3 +1,4 @@
---
apiVersion: apps/v1 apiVersion: apps/v1
kind: Deployment kind: Deployment
metadata: metadata:
@@ -22,7 +23,7 @@ spec:
kubernetes.io/hostname: home.homenet kubernetes.io/hostname: home.homenet
containers: containers:
- name: desubot - name: desubot
image: "ultradesu/desubot:latest" image: 'ultradesu/desubot:latest'
imagePullPolicy: Always imagePullPolicy: Always
envFrom: envFrom:
- secretRef: - secretRef:
@@ -31,11 +32,11 @@ spec:
- name: RUST_LOG - name: RUST_LOG
value: "info" value: "info"
volumeMounts: volumeMounts:
- mountPath: /storage - mountPath: /storage
name: storage name: storage
volumes: volumes:
- name: storage - name: storage
persistentVolumeClaim: nfs:
claimName: desubot-storage server: nas.homenet
readOnly: false path: /mnt/storage/Storage/k8s/desubot/
readOnly: false

View File

@@ -30,7 +30,7 @@ spec:
name: get-id-bot name: get-id-bot
env: env:
- name: RUST_LOG - name: RUST_LOG
value: "info,teloxide::error_handlers=off" value: "info"

View File

@@ -7,6 +7,3 @@ resources:
- get-id-bot.yaml - get-id-bot.yaml
- external-secrets.yaml - external-secrets.yaml
- desubot.yaml - desubot.yaml
- restart-job.yaml
- storage.yaml

View File

@@ -1,56 +0,0 @@
---
apiVersion: v1
kind: ServiceAccount
metadata:
name: tg-bots-restart-sa
---
apiVersion: rbac.authorization.k8s.io/v1
kind: Role
metadata:
name: tg-bots-restart-role
rules:
- apiGroups: ["apps"]
resources: ["deployments"]
verbs: ["get", "patch"]
---
apiVersion: rbac.authorization.k8s.io/v1
kind: RoleBinding
metadata:
name: tg-bots-restart-rb
subjects:
- kind: ServiceAccount
name: tg-bots-restart-sa
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: Role
name: tg-bots-restart-role
---
apiVersion: batch/v1
kind: CronJob
metadata:
name: tg-bots-daily-restart
spec:
schedule: "0 4 * * *" # every day at 04:00
jobTemplate:
spec:
template:
spec:
serviceAccountName: tg-bots-restart-sa
restartPolicy: OnFailure
containers:
- name: kubectl
image: bitnami/kubectl:latest
env:
- name: POD_NAMESPACE
valueFrom:
fieldRef:
fieldPath: metadata.namespace
command:
- /bin/sh
- -c
- |
kubectl -n "$POD_NAMESPACE" rollout restart deployment/desubot
kubectl -n "$POD_NAMESPACE" rollout restart deployment/get-id-bot

View File

@@ -1,12 +0,0 @@
---
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
name: desubot-storage
spec:
accessModes:
- ReadWriteMany
storageClassName: nfs-csi
resources:
requests:
storage: 200Gi

View File

@@ -1,21 +0,0 @@
apiVersion: argoproj.io/v1alpha1
kind: Application
metadata:
name: xandikos
namespace: argocd
spec:
project: apps
destination:
namespace: xandikos
server: https://kubernetes.default.svc
source:
repoURL: ssh://git@gt.hexor.cy:30022/ab/homelab.git
targetRevision: HEAD
path: k8s/apps/xandikos
syncPolicy:
automated:
selfHeal: true
prune: true
syncOptions:
- CreateNamespace=true

View File

@@ -1,70 +0,0 @@
---
apiVersion: apps/v1
kind: Deployment
metadata:
name: xandikos
labels:
app: xandikos
spec:
selector:
matchLabels:
app: xandikos
replicas: 1
strategy:
type: RollingUpdate
rollingUpdate:
maxSurge: 1
maxUnavailable: 0
template:
metadata:
labels:
app: xandikos
spec:
nodeSelector:
kubernetes.io/hostname: master.tail2fe2d.ts.net
volumes:
- name: storage
hostPath:
path: /k8s/xandikos
type: Directory
containers:
- name: xandikos
image: ghcr.io/jelmer/xandikos:latest
imagePullPolicy: Always
command:
- "python3"
- "-m"
- "xandikos.web"
- "--port=8081"
- "-d/data"
- "--defaults"
- "--listen-address=0.0.0.0"
- "--route-prefix=/dav"
resources:
requests:
memory: "64Mi"
cpu: "100m"
limits:
memory: "512Mi"
cpu: "1000m"
livenessProbe:
httpGet:
path: /
port: 8081
initialDelaySeconds: 30
periodSeconds: 10
timeoutSeconds: 5
readinessProbe:
httpGet:
path: /
port: 8081
initialDelaySeconds: 10
periodSeconds: 5
timeoutSeconds: 3
ports:
- name: http
containerPort: 8081
protocol: TCP
volumeMounts:
- name: storage
mountPath: /data

View File

@@ -1,31 +0,0 @@
---
apiVersion: external-secrets.io/v1
kind: ExternalSecret
metadata:
name: mmdl-secrets
spec:
target:
name: mmdl-secrets
deletionPolicy: Delete
template:
type: Opaque
data:
DB_DIALECT: 'postgres'
DB_HOST: psql.psql.svc
DB_USER: mmdl
DB_NAME: mmdl
DB_PORT: "5432"
DB_PASS: |-
{{ .pg_pass }}
AES_PASSWORD: |-
{{ .pg_pass }}
data:
- secretKey: pg_pass
sourceRef:
storeRef:
name: vaultwarden-login
kind: ClusterSecretStore
remoteRef:
key: 2a9deb39-ef22-433e-a1be-df1555625e22
property: fields[12].value

View File

@@ -1,47 +0,0 @@
---
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:
name: xandikos
annotations:
ingressClassName: traefik
cert-manager.io/cluster-issuer: letsencrypt
traefik.ingress.kubernetes.io/router.middlewares: kube-system-https-redirect@kubernetescrd
acme.cert-manager.io/http01-edit-in-place: "true"
spec:
rules:
- host: cal.hexor.cy
http:
paths:
- path: /
pathType: Prefix
backend:
service:
name: mmdl
port:
number: 3000
- path: /dav
pathType: Prefix
backend:
service:
name: xandikos
port:
number: 8081
- path: /.well-known/carddav
pathType: Exact
backend:
service:
name: xandikos
port:
number: 8081
- path: /.well-known/caldav
pathType: Exact
backend:
service:
name: xandikos
port:
number: 8081
tls:
- secretName: xandikos-tls
hosts:
- cal.hexor.cy

View File

@@ -1,11 +0,0 @@
apiVersion: kustomize.config.k8s.io/v1beta1
kind: Kustomization
resources:
- deployment.yaml
- service.yaml
- mmdl-deployment.yaml
- mmdl-service.yaml
- ingress.yaml
- external-secrets.yaml

View File

@@ -1,61 +0,0 @@
---
apiVersion: apps/v1
kind: Deployment
metadata:
name: mmdl
labels:
app: mmdl
spec:
selector:
matchLabels:
app: mmdl
replicas: 1
strategy:
type: RollingUpdate
rollingUpdate:
maxSurge: 1
maxUnavailable: 0
template:
metadata:
labels:
app: mmdl
spec:
nodeSelector:
kubernetes.io/hostname: master.tail2fe2d.ts.net
containers:
- name: mmdl
image: intriin/mmdl:latest
imagePullPolicy: Always
envFrom:
- secretRef:
name: mmdl-secrets
env:
- name: NEXTAUTH_URL
value: "https://cal.hexor.cy"
- name: CALDAV_SERVER_URL
value: "https://cal.hexor.cy/dav"
resources:
requests:
memory: "128Mi"
cpu: "100m"
limits:
memory: "512Mi"
cpu: "1000m"
livenessProbe:
httpGet:
path: /
port: 3000
initialDelaySeconds: 30
periodSeconds: 10
timeoutSeconds: 5
readinessProbe:
httpGet:
path: /
port: 3000
initialDelaySeconds: 10
periodSeconds: 5
timeoutSeconds: 3
ports:
- name: http
containerPort: 3000
protocol: TCP

View File

@@ -1,14 +0,0 @@
---
apiVersion: v1
kind: Service
metadata:
name: mmdl
spec:
selector:
app: mmdl
type: ClusterIP
ports:
- name: http
port: 3000
protocol: TCP
targetPort: 3000

View File

@@ -1,16 +0,0 @@
---
apiVersion: v1
kind: Service
metadata:
name: xandikos
labels:
app: xandikos
spec:
selector:
app: xandikos
ports:
- protocol: TCP
port: 8081
targetPort: 8081
name: http
type: ClusterIP

View File

@@ -47,20 +47,3 @@ spec:
server: https://kubernetes.default.svc server: https://kubernetes.default.svc
sourceRepos: sourceRepos:
- ssh://git@gt.hexor.cy:30022/ab/homelab.git - ssh://git@gt.hexor.cy:30022/ab/homelab.git
---
apiVersion: argoproj.io/v1alpha1
kind: AppProject
metadata:
name: desktop
namespace: argocd
spec:
clusterResourceWhitelist:
- group: '*'
kind: '*'
description: Hexor Home Lab Desktop Apps
destinations:
- namespace: '*'
server: https://kubernetes.default.svc
sourceRepos:
- ssh://git@gt.hexor.cy:30022/ab/homelab.git

View File

@@ -2,7 +2,7 @@
global: global:
domain: ag.hexor.cy domain: ag.hexor.cy
nodeSelector: &nodeSelector nodeSelector:
kubernetes.io/hostname: master.tail2fe2d.ts.net kubernetes.io/hostname: master.tail2fe2d.ts.net
logging: logging:
format: text format: text
@@ -55,15 +55,15 @@ configs:
controller: controller:
replicas: 1 replicas: 1
nodeSelector: nodeSelector:
<<: *nodeSelector kubernetes.io/hostname: master.tail2fe2d.ts.net
# Add resources (requests/limits), PDB etc. if needed # Add resources (requests/limits), PDB etc. if needed
# Dex OIDC provider # Dex OIDC provider
dex: dex:
replicas: 1 replicas: 1
nodeSelector: nodeSelector:
<<: *nodeSelector kubernetes.io/hostname: master.tail2fe2d.ts.net
enabled: false enabled: false
# Standard Redis disabled because Redis HA is enabled # Standard Redis disabled because Redis HA is enabled
@@ -86,7 +86,7 @@ redis-ha:
server: server:
replicas: 1 replicas: 1
nodeSelector: nodeSelector:
<<: *nodeSelector kubernetes.io/hostname: master.tail2fe2d.ts.net
ingress: ingress:
enabled: false enabled: false
@@ -99,11 +99,8 @@ server:
# Repository Server # Repository Server
repoServer: repoServer:
replicas: 1 replicas: 1
livenessProbe:
timeoutSeconds: 10
periodSeconds: 60
nodeSelector: nodeSelector:
<<: *nodeSelector kubernetes.io/hostname: master.tail2fe2d.ts.net
# Add resources (requests/limits), PDB etc. if needed # Add resources (requests/limits), PDB etc. if needed
# ApplicationSet Controller # ApplicationSet Controller
@@ -111,7 +108,7 @@ applicationSet:
enabled: true # Enabled by default enabled: true # Enabled by default
replicas: 1 replicas: 1
nodeSelector: nodeSelector:
<<: *nodeSelector kubernetes.io/hostname: master.tail2fe2d.ts.net
# Add resources (requests/limits), PDB etc. if needed # Add resources (requests/limits), PDB etc. if needed
# Notifications Controller # Notifications Controller
@@ -119,5 +116,5 @@ notifications:
enabled: true # Enabled by default enabled: true # Enabled by default
replicas: 1 replicas: 1
nodeSelector: nodeSelector:
<<: *nodeSelector kubernetes.io/hostname: master.tail2fe2d.ts.net
# Add notifiers, triggers, templates configurations if needed # Add notifiers, triggers, templates configurations if needed

View File

@@ -35,6 +35,5 @@ spec:
key: secretKey key: secretKey
selector: selector:
dnsZones: dnsZones:
- "ps.hexor.cy"
- "of.hexor.cy" - "of.hexor.cy"

View File

@@ -1,6 +1,6 @@
FROM debian:sid FROM debian:sid
ENV BW_CLI_VERSION=2025.12.1 ENV BW_CLI_VERSION=2025.5.0
RUN apt update && \ RUN apt update && \
apt install -y wget unzip && \ apt install -y wget unzip && \

View File

@@ -37,15 +37,15 @@ spec:
kubernetes.io/hostname: master.tail2fe2d.ts.net kubernetes.io/hostname: master.tail2fe2d.ts.net
containers: containers:
- name: bitwarden-cli - name: bitwarden-cli
image: ultradesu/bitwarden-client:2025.12.1 image: ultradesu/bitwarden-client:2025.5.0
imagePullPolicy: Always imagePullPolicy: Always
resources: resources:
requests: requests:
memory: "128Mi" memory: "128Mi"
cpu: "300m" cpu: "100m"
limits: limits:
memory: "512Mi" memory: "512Mi"
cpu: "1000m" cpu: "500m"
env: env:
- name: BW_HOST - name: BW_HOST
valueFrom: valueFrom:

View File

@@ -3,15 +3,5 @@ kind: Kustomization
resources: resources:
- app.yaml - app.yaml
- nfs-storage.yaml
- coredns-internal-resolve.yaml - coredns-internal-resolve.yaml
helmCharts:
- name: csi-driver-nfs
repo: https://raw.githubusercontent.com/kubernetes-csi/csi-driver-nfs/master/charts
version: 4.12.0
releaseName: csi-driver-nfs
namespace: kube-system
#valuesFile: values.yaml
includeCRDs: true

View File

@@ -1,14 +0,0 @@
---
apiVersion: storage.k8s.io/v1
kind: StorageClass
metadata:
name: nfs-csi
provisioner: nfs.csi.k8s.io
parameters:
server: nas.homenet
share: /mnt/storage/Storage/PVC
reclaimPolicy: Retain
volumeBindingMode: Immediate
mountOptions:
- vers=4
- hard

View File

@@ -121,12 +121,6 @@ spec:
{{ .pasarguard }} {{ .pasarguard }}
USER_remnawave: |- USER_remnawave: |-
{{ .remnawave }} {{ .remnawave }}
USER_umami: |-
{{ .umami }}
USER_mmdl: |-
{{ .mmdl }}
USER_n8n: |-
{{ .n8n }}
data: data:
- secretKey: authentik - secretKey: authentik
sourceRef: sourceRef:
@@ -238,37 +232,3 @@ spec:
metadataPolicy: None metadataPolicy: None
key: 2a9deb39-ef22-433e-a1be-df1555625e22 key: 2a9deb39-ef22-433e-a1be-df1555625e22
property: fields[10].value property: fields[10].value
- secretKey: umami
sourceRef:
storeRef:
name: vaultwarden-login
kind: ClusterSecretStore
remoteRef:
conversionStrategy: Default
decodingStrategy: None
metadataPolicy: None
key: 2a9deb39-ef22-433e-a1be-df1555625e22
property: fields[11].value
- secretKey: mmdl
sourceRef:
storeRef:
name: vaultwarden-login
kind: ClusterSecretStore
remoteRef:
conversionStrategy: Default
decodingStrategy: None
metadataPolicy: None
key: 2a9deb39-ef22-433e-a1be-df1555625e22
property: fields[12].value
- secretKey: n8n
sourceRef:
storeRef:
name: vaultwarden-login
kind: ClusterSecretStore
remoteRef:
conversionStrategy: Default
decodingStrategy: None
metadataPolicy: None
key: 2a9deb39-ef22-433e-a1be-df1555625e22
property: fields[13].value

View File

@@ -28,7 +28,7 @@ spec:
type: DirectoryOrCreate type: DirectoryOrCreate
containers: containers:
- name: psql - name: psql
image: 'bitnamilegacy/postgresql:17' image: 'bitnami/postgresql:17'
env: env:
- name: POSTGRESQL_PASSWORD - name: POSTGRESQL_PASSWORD
valueFrom: valueFrom:
@@ -63,7 +63,7 @@ spec:
containerPort: 9187 containerPort: 9187
protocol: TCP protocol: TCP
- name: user-creation - name: user-creation
image: 'bitnamilegacy/postgresql:17' image: 'bitnami/postgresql:17'
command: command:
- /bin/bash - /bin/bash
- -c - -c

View File

@@ -28,7 +28,7 @@ spec:
type: DirectoryOrCreate type: DirectoryOrCreate
containers: containers:
- name: psql - name: psql
image: 'bitnamilegacy/postgresql:17' image: 'bitnami/postgresql:17'
env: env:
- name: POSTGRESQL_PASSWORD - name: POSTGRESQL_PASSWORD
valueFrom: valueFrom:
@@ -60,7 +60,7 @@ spec:
containerPort: 9187 containerPort: 9187
protocol: TCP protocol: TCP
- name: user-creation - name: user-creation
image: 'bitnamilegacy/postgresql:17' image: 'bitnami/postgresql:17'
command: command:
- /bin/bash - /bin/bash
- -c - -c

View File

@@ -1,5 +1,5 @@
image: image:
tag: "latest" tag: "9.9"
pullPolicy: Always pullPolicy: Always
env: env:
email: "postgres@hexor.cy" email: "postgres@hexor.cy"

View File

@@ -13,6 +13,9 @@ spec:
targetRevision: HEAD targetRevision: HEAD
path: k8s/core/prom-stack path: k8s/core/prom-stack
syncPolicy: syncPolicy:
automated:
selfHeal: true
prune: true
syncOptions: syncOptions:
- CreateNamespace=true - CreateNamespace=true
- ServerSideApply=true - ServerSideApply=true

View File

@@ -79,83 +79,3 @@ spec:
key: 2a9deb39-ef22-433e-a1be-df1555625e22 key: 2a9deb39-ef22-433e-a1be-df1555625e22
property: fields[2].value property: fields[2].value
---
apiVersion: external-secrets.io/v1
kind: ExternalSecret
metadata:
name: alertmanager-telegram
spec:
target:
name: alertmanager-telegram-secret
deletionPolicy: Delete
template:
type: Opaque
data:
TELEGRAM_BOT_TOKEN: |-
{{ .bot_token }}
TELEGRAM_CHAT_ID: |-
{{ .chat_id }}
data:
- secretKey: bot_token
sourceRef:
storeRef:
name: vaultwarden-login
kind: ClusterSecretStore
remoteRef:
conversionStrategy: Default
decodingStrategy: None
metadataPolicy: None
key: eca0fb0b-3939-40a8-890a-6294863e5a65
property: fields[0].value
- secretKey: chat_id
sourceRef:
storeRef:
name: vaultwarden-login
kind: ClusterSecretStore
remoteRef:
conversionStrategy: Default
decodingStrategy: None
metadataPolicy: None
key: eca0fb0b-3939-40a8-890a-6294863e5a65
property: fields[1].value
---
apiVersion: external-secrets.io/v1
kind: ExternalSecret
metadata:
name: grafana-telegram
spec:
target:
name: grafana-telegram
deletionPolicy: Delete
template:
type: Opaque
data:
bot-token: |-
{{ .bot_token }}
chat-id: |-
{{ .chat_id }}
data:
- secretKey: bot_token
sourceRef:
storeRef:
name: vaultwarden-login
kind: ClusterSecretStore
remoteRef:
conversionStrategy: Default
decodingStrategy: None
metadataPolicy: None
key: eca0fb0b-3939-40a8-890a-6294863e5a65
property: fields[0].value
- secretKey: chat_id
sourceRef:
storeRef:
name: vaultwarden-login
kind: ClusterSecretStore
remoteRef:
conversionStrategy: Default
decodingStrategy: None
metadataPolicy: None
key: eca0fb0b-3939-40a8-890a-6294863e5a65
property: fields[1].value

View File

@@ -1,152 +0,0 @@
apiVersion: v1
kind: ConfigMap
metadata:
name: grafana-alerting
namespace: prometheus
data:
rules.yaml: |
apiVersion: 1
groups:
- orgId: 1
name: pasarguard_alerts
folder: Kubernetes
interval: 1m
rules:
- uid: pasarguard_cpu_throttling
title: VPN CPU Throttle
condition: B
data:
- refId: A
relativeTimeRange:
from: 600
to: 0
datasourceUid: P76F38748CEC837F0
model:
expr: 'rate(container_cpu_cfs_throttled_periods_total{container="pasarguard-node"}[5m])'
refId: A
intervalMs: 1000
maxDataPoints: 43200
- refId: B
relativeTimeRange:
from: 600
to: 0
datasourceUid: __expr__
model:
conditions:
- evaluator:
params:
- 0.1
type: gt
operator:
type: and
query:
params: []
datasource:
type: __expr__
uid: __expr__
expression: A
reducer: last
refId: B
type: reduce
noDataState: NoData
execErrState: Alerting
for: 5m
annotations:
pod: '{{ $labels.pod }}'
node: '{{ $labels.node }}'
namespace: '{{ $labels.namespace }}'
throttle_rate: '{{ printf "%.2f" $values.A }}'
summary: 'VPN node throttling CPU'
labels:
severity: warning
- orgId: 1
name: kubernetes_alerts
folder: Kubernetes
interval: 30s
rules:
- uid: node_not_ready
title: Kubernetes Node Not Ready
condition: B
data:
- refId: A
relativeTimeRange:
from: 300
to: 0
datasourceUid: P76F38748CEC837F0
model:
expr: 'kube_node_status_condition{condition="Ready",status="true"} == 0'
refId: A
intervalMs: 1000
maxDataPoints: 43200
- refId: B
relativeTimeRange:
from: 300
to: 0
datasourceUid: __expr__
model:
conditions:
- evaluator:
params:
- 0
type: gt
operator:
type: and
query:
params: []
datasource:
type: __expr__
uid: __expr__
expression: A
reducer: last
refId: B
type: reduce
noDataState: Alerting
execErrState: Alerting
for: 0s
annotations:
node: '{{ $labels.node }}'
condition: '{{ $labels.condition }}'
summary: 'Kubernetes node is not ready'
labels:
severity: critical
contactpoints.yaml: |
apiVersion: 1
contactPoints:
- orgId: 1
name: telegram
receivers:
- uid: telegram_default
type: telegram
disableResolveMessage: false
settings:
bottoken: $TELEGRAM_BOT_TOKEN
chatid: "124317807"
message: |
{{ if eq .Status "firing" }}🔥 FIRING{{ else }}✅ RESOLVED{{ end }}
{{ range .Alerts }}
📊 <b>{{ .Labels.alertname }}</b>
{{ .Annotations.summary }}
{{ if .Annotations.node }}🖥 <b>Node:</b> <code>{{ .Annotations.node }}</code>{{ end }}
{{ if .Annotations.pod }}📦 <b>Pod:</b> <code>{{ .Annotations.pod }}</code>{{ end }}
{{ if .Annotations.namespace }}📁 <b>Namespace:</b> <code>{{ .Annotations.namespace }}</code>{{ end }}
{{ if .Annotations.throttle_rate }}⚠️ <b>Throttling rate:</b> {{ .Annotations.throttle_rate }}{{ end }}
🔗 <a href="{{ .GeneratorURL }}">View in Grafana</a>
{{ end }}
parse_mode: HTML
policies.yaml: |
apiVersion: 1
policies:
- orgId: 1
receiver: telegram
group_by:
- grafana_folder
- alertname
group_wait: 10s
group_interval: 5m
repeat_interval: 4h

View File

@@ -38,10 +38,6 @@ datasources:
url: http://prometheus-kube-prometheus-prometheus.prometheus.svc:9090 url: http://prometheus-kube-prometheus-prometheus.prometheus.svc:9090
access: proxy access: proxy
isDefault: true isDefault: true
- name: Loki
type: loki
url: http://loki-gateway.prometheus.svc:80
access: proxy
ingress: ingress:
enabled: true enabled: true
@@ -56,30 +52,3 @@ ingress:
hosts: hosts:
- '*.hexor.cy' - '*.hexor.cy'
extraConfigmapMounts:
- name: grafana-alerting-rules
mountPath: /etc/grafana/provisioning/alerting/rules.yaml
configMap: grafana-alerting
subPath: rules.yaml
readOnly: true
- name: grafana-alerting-contactpoints
mountPath: /etc/grafana/provisioning/alerting/contactpoints.yaml
configMap: grafana-alerting
subPath: contactpoints.yaml
readOnly: true
- name: grafana-alerting-policies
mountPath: /etc/grafana/provisioning/alerting/policies.yaml
configMap: grafana-alerting
subPath: policies.yaml
readOnly: true
envValueFrom:
TELEGRAM_BOT_TOKEN:
secretKeyRef:
name: grafana-telegram
key: bot-token
TELEGRAM_CHAT_ID:
secretKeyRef:
name: grafana-telegram
key: chat-id

View File

@@ -2,9 +2,9 @@ apiVersion: kustomize.config.k8s.io/v1beta1
kind: Kustomization kind: Kustomization
resources: resources:
- app.yaml
- persistentVolume.yaml - persistentVolume.yaml
- external-secrets.yaml - external-secrets.yaml
- grafana-alerting-configmap.yaml
helmCharts: helmCharts:
- name: kube-prometheus-stack - name: kube-prometheus-stack
@@ -23,18 +23,3 @@ helmCharts:
valuesFile: grafana-values.yaml valuesFile: grafana-values.yaml
includeCRDs: true includeCRDs: true
- name: loki
repo: https://grafana.github.io/helm-charts
version: 6.29.0
releaseName: loki
namespace: prometheus
valuesFile: loki-values.yaml
includeCRDs: true
- name: promtail
repo: https://grafana.github.io/helm-charts
version: 6.16.6
releaseName: promtail
namespace: prometheus
valuesFile: promtail-values.yaml

View File

@@ -1,75 +0,0 @@
# Loki SingleBinary mode - optimal for homelab
deploymentMode: SingleBinary
loki:
auth_enabled: false
commonConfig:
replication_factor: 1
path_prefix: /var/loki
schemaConfig:
configs:
- from: 2024-01-01
store: tsdb
object_store: filesystem
schema: v13
index:
prefix: index_
period: 24h
storage:
type: filesystem
filesystem:
chunks_directory: /var/loki/chunks
rules_directory: /var/loki/rules
limits_config:
reject_old_samples: false
ingestion_rate_mb: 16
ingestion_burst_size_mb: 32
max_query_parallelism: 32
volume_enabled: true
singleBinary:
replicas: 1
nodeSelector:
kubernetes.io/hostname: master.tail2fe2d.ts.net
persistence:
enabled: true
size: 50Gi
storageClass: ""
# Disable distributed mode components
read:
replicas: 0
write:
replicas: 0
backend:
replicas: 0
# Disable memcached (not needed for SingleBinary)
chunksCache:
enabled: false
resultsCache:
enabled: false
# Gateway for Loki access
gateway:
enabled: true
replicas: 1
service:
type: ClusterIP
# Disable tests and canary
test:
enabled: false
lokiCanary:
enabled: false
# Monitoring
monitoring:
dashboards:
enabled: false
rules:
enabled: false
serviceMonitor:
enabled: false
selfMonitoring:
enabled: false

View File

@@ -1,35 +1,5 @@
grafana: grafana:
enabled: false enabled: false
alertmanager:
config:
global:
telegram_api_url: "https://api.telegram.org"
route:
group_by: ['alertname', 'cluster', 'service']
group_wait: 10s
group_interval: 10s
repeat_interval: 12h
receiver: 'telegram'
receivers:
- name: 'telegram'
telegram_configs:
- bot_token: '${TELEGRAM_BOT_TOKEN}'
chat_id: ${TELEGRAM_CHAT_ID}
parse_mode: 'HTML'
message: |
{{ range .Alerts }}
<b>{{ .Labels.alertname }}</b>
{{ if .Labels.severity }}<b>Severity:</b> {{ .Labels.severity }}{{ end }}
<b>Status:</b> {{ .Status }}
{{ if .Annotations.summary }}<b>Summary:</b> {{ .Annotations.summary }}{{ end }}
{{ if .Annotations.description }}<b>Description:</b> {{ .Annotations.description }}{{ end }}
{{ end }}
alertmanagerSpec:
secrets:
- alertmanager-telegram-secret
prometheus: prometheus:
prometheusSpec: prometheusSpec:
enableRemoteWriteReceiver: true enableRemoteWriteReceiver: true

View File

@@ -1,37 +0,0 @@
# Promtail - log collection agent for all cluster pods
config:
clients:
- url: http://loki-gateway.prometheus.svc:80/loki/api/v1/push
# DaemonSet - runs on every node
daemonset:
enabled: true
# Tolerations for master/control-plane nodes
tolerations:
- key: node-role.kubernetes.io/master
operator: Exists
effect: NoSchedule
- key: node-role.kubernetes.io/control-plane
operator: Exists
effect: NoSchedule
# Init container to increase inotify limits
initContainer:
- name: init-inotify
image: docker.io/busybox:1.36
imagePullPolicy: IfNotPresent
command:
- sh
- -c
- sysctl -w fs.inotify.max_user_instances=512
securityContext:
privileged: true
resources:
requests:
cpu: 50m
memory: 64Mi
limits:
cpu: 200m
memory: 128Mi

View File

@@ -16,7 +16,7 @@ spec:
serviceAccountName: system-upgrade serviceAccountName: system-upgrade
upgrade: upgrade:
image: rancher/k3s-upgrade image: rancher/k3s-upgrade
version: v1.34.3+k3s1 version: v1.34.1+k3s1
--- ---
# Agent plan # Agent plan
apiVersion: upgrade.cattle.io/v1 apiVersion: upgrade.cattle.io/v1
@@ -39,5 +39,5 @@ spec:
serviceAccountName: system-upgrade serviceAccountName: system-upgrade
upgrade: upgrade:
image: rancher/k3s-upgrade image: rancher/k3s-upgrade
version: v1.34.3+k3s1 version: v1.34.1+k3s1

View File

@@ -1,21 +0,0 @@
apiVersion: argoproj.io/v1alpha1
kind: Application
metadata:
name: jellyfin-uk
namespace: argocd
spec:
project: apps
destination:
namespace: jellyfin-uk
server: https://kubernetes.default.svc
source:
repoURL: ssh://git@gt.hexor.cy:30022/ab/homelab.git
targetRevision: HEAD
path: k8s/desktop/jellyfin
syncPolicy:
automated:
selfHeal: true
prune: true
syncOptions:
- CreateNamespace=true

View File

@@ -1,16 +0,0 @@
apiVersion: kustomize.config.k8s.io/v1beta1
kind: Kustomization
resources:
- app.yaml
- qbittorent.yaml
helmCharts:
- name: jellyfin
repo: https://utkuozdemir.org/helm-charts
version: 2.0.0
releaseName: jellyfin
namespace: jellyfin
valuesFile: values.yaml
includeCRDs: true

View File

@@ -1,123 +0,0 @@
---
apiVersion: external-secrets.io/v1
kind: ExternalSecret
metadata:
name: vpn-creds
spec:
target:
name: vpn-creds
deletionPolicy: Delete
template:
type: Opaque
data:
ss_link: |-
{{ .ss_link }}
data:
- secretKey: ss_link
sourceRef:
storeRef:
name: vaultwarden-login
kind: ClusterSecretStore
remoteRef:
key: cfee6f62-fb06-4a4c-b6d8-92da4908c65a
property: fields[0].value
---
apiVersion: apps/v1
kind: Deployment
metadata:
name: qbittorrent
labels:
app: qbittorrent
annotations:
reloader.stakater.com/auto: "true"
spec:
selector:
matchLabels:
app: qbittorrent
replicas: 1
strategy:
type: RollingUpdate
rollingUpdate:
maxSurge: 1
maxUnavailable: 0
template:
metadata:
labels:
app: qbittorrent
spec:
nodeSelector:
kubernetes.io/hostname: uk-desktop.tail2fe2d.ts.net
tolerations:
- key: workload
operator: Equal
value: desktop
effect: NoSchedule
volumes:
- name: config
hostPath:
path: /k8s/qbt-config
type: DirectoryOrCreate
- name: media
hostPath:
path: /k8s/media/downloads
type: DirectoryOrCreate
containers:
- name: qbittorrent
image: 'linuxserver/qbittorrent:latest'
ports:
- name: http
containerPort: 8080
protocol: TCP
volumeMounts:
- name: config
mountPath: /config
- name: media
mountPath: /downloads
- name: shadowsocks-proxy
image: teddysun/shadowsocks-rust:latest
env:
- name: SS_LINK
valueFrom:
secretKeyRef:
name: vpn-creds
key: ss_link
command: ["/bin/bash", "-c", "rm /etc/shadowsocks-rust/config.json && sslocal --server-url $SS_LINK --local-addr 127.0.0.1:8081 -U --protocol http"]
resources:
requests:
memory: "64Mi"
cpu: "300m"
limits:
memory: "128Mi"
cpu: "300m"
---
apiVersion: v1
kind: Service
metadata:
name: qbittorrent
spec:
selector:
app: qbittorrent
ports:
- protocol: TCP
port: 80
targetPort: 8080
---
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:
name: jf-local-ingress
annotations:
ingressClassName: traefik
spec:
rules:
- host: tr.uk
http:
paths:
- path: /
pathType: Prefix
backend:
service:
name: qbittorrent
port:
number: 80

View File

@@ -1,41 +0,0 @@
image:
tag: 10.11.4
resources:
requests:
memory: "2Gi"
cpu: "1000m"
limits:
memory: "8Gi"
cpu: "6000m"
nodeSelector:
kubernetes.io/hostname: uk-desktop.tail2fe2d.ts.net
tolerations:
- key: workload
operator: Equal
value: desktop
effect: NoSchedule
persistence:
config:
enabled: true
isPvc: false
customVolume:
hostPath:
path: /k8s/jellyfin
type: DirectoryOrCreate
data:
enabled: true
isPvc: false
customVolume:
hostPath:
path: /k8s/media/downloads
type: DirectoryOrCreate
ingress:
enabled: true
className: traefik
hosts:
- host: jf.uk
paths:
- path: /
pathType: Prefix

View File

@@ -1,18 +0,0 @@
apiVersion: argoproj.io/v1alpha1
kind: Application
metadata:
name: khm-client
namespace: argocd
spec:
project: desktop
destination:
namespace: khm
server: https://kubernetes.default.svc
source:
repoURL: ssh://git@gt.hexor.cy:30022/ab/homelab.git
targetRevision: HEAD
path: k8s/desktop/khm
syncPolicy:
automated:
selfHeal: true
prune: true

View File

@@ -1,33 +0,0 @@
---
apiVersion: external-secrets.io/v1
kind: ExternalSecret
metadata:
name: khm-client-creds
spec:
target:
name: khm-client-creds
deletionPolicy: Delete
template:
type: Opaque
data:
USERNAME: |-
{{ .username }}
PASSWORD: |-
{{ .password }}
data:
- secretKey: username
sourceRef:
storeRef:
name: vaultwarden-login
kind: ClusterSecretStore
remoteRef:
key: 19c06480-0814-4d1f-aa80-710105989188
property: login.username
- secretKey: password
sourceRef:
storeRef:
name: vaultwarden-login
kind: ClusterSecretStore
remoteRef:
key: 19c06480-0814-4d1f-aa80-710105989188
property: login.password

View File

@@ -1,69 +0,0 @@
---
apiVersion: batch/v1
kind: CronJob
metadata:
name: khm-client
labels:
app: khm-client
spec:
schedule: "15 * * * *"
concurrencyPolicy: Forbid
successfulJobsHistoryLimit: 3
failedJobsHistoryLimit: 3
jobTemplate:
spec:
template:
metadata:
labels:
app: khm-client
spec:
restartPolicy: OnFailure
hostNetwork: true
nodeSelector:
node-role.kubernetes.io/desktop: ""
tolerations:
- key: workload
operator: Equal
value: desktop
effect: NoSchedule
containers:
- name: khm-client
image: 'ultradesu/khm:latest'
imagePullPolicy: Always
securityContext:
privileged: false
resources:
requests:
memory: "64Mi"
cpu: "50m"
limits:
memory: "256Mi"
cpu: "200m"
command:
- /bin/sh
- -c
- |
/usr/local/bin/khm \
--known-hosts /host-ssh/known_hosts \
--host https://khm.hexor.cy \
--flow=private \
--basic-auth="${USERNAME}:${PASSWORD}" \
--in-place
env:
- name: USERNAME
valueFrom:
secretKeyRef:
name: khm-client-creds
key: USERNAME
- name: PASSWORD
valueFrom:
secretKeyRef:
name: khm-client-creds
key: PASSWORD
volumeMounts:
- name: known-hosts
mountPath: /host-ssh/known_hosts
volumes:
- name: known-hosts
hostPath:
path: /home/ab/.ssh/known_hosts

View File

@@ -1,6 +0,0 @@
apiVersion: kustomize.config.k8s.io/v1beta1
kind: Kustomization
resources:
- external-secrets.yaml
- khm-client-cronjob.yaml

View File

@@ -1,3 +1,4 @@
data "authentik_flow" "default_authorization_flow" { data "authentik_flow" "default_authorization_flow" {
slug = var.default_authorization_flow slug = var.default_authorization_flow
} }
@@ -298,7 +299,7 @@ resource "authentik_outpost" "outposts" {
kubernetes_ingress_class_name = null kubernetes_ingress_class_name = null
kubernetes_disabled_components = [] kubernetes_disabled_components = []
kubernetes_ingress_annotations = {} kubernetes_ingress_annotations = {}
kubernetes_ingress_secret_name = "idm-tls" kubernetes_ingress_secret_name = "authentik-outpost-tls"
}) })
depends_on = [ depends_on = [

View File

@@ -8,6 +8,6 @@ terraform {
} }
provider "authentik" { provider "authentik" {
url = var.authentik_url url = "https://idm.hexor.cy"
token = var.authentik_token token = "qXcuoCg77JaRgqnU6rqIIBa8MBJ4UNyLPTL89dZI8zeC2jfaWqQ7k56BJs8F"
} }

View File

@@ -51,9 +51,6 @@ proxy_applications = {
internal_host = "http://secret-reader.k8s-secret.svc:80" internal_host = "http://secret-reader.k8s-secret.svc:80"
internal_host_ssl_validation = false internal_host_ssl_validation = false
meta_description = "" meta_description = ""
skip_path_regex = <<-EOT
/webhook
EOT
meta_icon = "https://img.icons8.com/ios-filled/50/password.png" meta_icon = "https://img.icons8.com/ios-filled/50/password.png"
mode = "proxy" mode = "proxy"
outpost = "kubernetes-outpost" outpost = "kubernetes-outpost"
@@ -132,12 +129,12 @@ EOT
access_groups = ["admins"] access_groups = ["admins"]
} }
"syncthing-router" = { "syncthing-master" = {
name = "Syncthing" name = "Syncthing-master"
slug = "syncthing" slug = "syncthing-master"
group = "Media and Storage" group = "Media and Storage"
external_host = "https://ss.hexor.cy" external_host = "https://ss.hexor.cy"
internal_host = "http://syncthing-router.syncthing.svc:80" internal_host = "https://syncthing-master.syncthing.svc:8384"
internal_host_ssl_validation = false internal_host_ssl_validation = false
meta_description = "" meta_description = ""
meta_icon = "https://img.icons8.com/?size=100&id=Id4NcEcXcYzF&format=png&color=000000" meta_icon = "https://img.icons8.com/?size=100&id=Id4NcEcXcYzF&format=png&color=000000"
@@ -199,10 +196,9 @@ EOT
internal_host_ssl_validation = false internal_host_ssl_validation = false
meta_description = "" meta_description = ""
skip_path_regex = <<-EOT skip_path_regex = <<-EOT
/
/sub/ /sub/
/dashboard/ /dashboard/
/api/ /
EOT EOT
meta_icon = "https://img.icons8.com/?size=100&id=fqAD3lAB6zTe&format=png&color=000000" meta_icon = "https://img.icons8.com/?size=100&id=fqAD3lAB6zTe&format=png&color=000000"
mode = "proxy" mode = "proxy"

View File

@@ -166,9 +166,3 @@ variable "authentik_url" {
type = string type = string
default = "https://idm.hexor.cy" default = "https://idm.hexor.cy"
} }
variable "authentik_token" {
description = "Authentik API token (set via TF_VAR_authentik_token env var)"
type = string
sensitive = true
}