Compare commits
87 Commits
auto-updat
...
auto-updat
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
2136cffa20 | ||
|
|
9df0a3c8b7 | ||
| 3157b0c325 | |||
|
|
cbe1c23709 | ||
|
|
70198ca1c2 | ||
|
|
39207fcb39 | ||
|
|
bae89f8738 | ||
|
|
72950dae83 | ||
| 9096b4bcf7 | |||
|
|
8b6b0a0cd6 | ||
| 9024a41a4c | |||
| 2b5e76e24d | |||
|
|
651acf665e | ||
| 7bd482a000 | |||
| 6a0be650ea | |||
| b78efdb81a | |||
| 7bf27506b0 | |||
| 5e3be29b7a | |||
| 44ce19b815 | |||
| 434b947896 | |||
| b13107882c | |||
| f76d44ce98 | |||
| 43c0fdf2f2 | |||
| 512eaf842e | |||
| 91cd967989 | |||
| d092401cd6 | |||
| 5b2768ad4f | |||
| a7d71e40a3 | |||
| 9cbaa23aed | |||
|
|
0b3fddbd40 | ||
|
|
ee0c55dcea | ||
|
|
f545a47162 | ||
|
|
fdbeb1a9a7 | ||
|
|
0fbeb96a6a | ||
|
|
0eba6143f4 | ||
|
|
b71f54f714 | ||
|
|
9a3bdfaf9c | ||
|
|
cfa275f4fc | ||
|
|
4a887b6775 | ||
| 981aa2ba15 | |||
| 872c0f4adf | |||
| 5b1ff26975 | |||
| 93bf782ece | |||
| f153bfc0b4 | |||
| 6b60fca39c | |||
| abb47a6db0 | |||
| e008ac3f59 | |||
| c945575ea1 | |||
| 01348dd99e | |||
| 0b4507a72d | |||
| 5cca64813a | |||
| 837094944e | |||
| 7da2fab580 | |||
| 007df29133 | |||
| b25a82ba1e | |||
| 5e7e9031a3 | |||
| 70ae7c9a50 | |||
| d95faaf2c1 | |||
| af7e29c063 | |||
| 4ea48f0f94 | |||
| 4bfc35d8e2 | |||
| 46c0fab78a | |||
| 6dc43149f4 | |||
| ca1efe6230 | |||
| e90d2c9dc5 | |||
| a884c2b969 | |||
| db92976872 | |||
|
|
d924ebd3ee | ||
|
|
4b30185655 | ||
|
|
a65b37f000 | ||
|
|
f394b4f9da | ||
| 5d12fc854a | |||
|
|
f415e0711e | ||
| 14dc69904c | |||
| f6dc7aa6e3 | |||
| badd82f9af | |||
| a5cb49471a | |||
| 79c23e14b0 | |||
| 5bc44e45b0 | |||
|
|
4a80f2f596 | ||
| b58461232c | |||
| be6e601275 | |||
| 063a4a502b | |||
|
|
22382b63a1 | ||
|
|
718709115f | ||
|
|
df78728137 | ||
|
|
9671dc3aa8 |
@@ -22,12 +22,13 @@ jobs:
|
|||||||
|
|
||||||
- name: Install Python dependencies
|
- name: Install Python dependencies
|
||||||
run: |
|
run: |
|
||||||
pip install pyyaml
|
python3 -m venv .venv
|
||||||
|
.venv/bin/pip install pyyaml
|
||||||
|
|
||||||
- name: Generate K8s Services Wiki
|
- name: Generate K8s Services Wiki
|
||||||
run: |
|
run: |
|
||||||
echo "📋 Starting K8s wiki generation..."
|
echo "📋 Starting K8s wiki generation..."
|
||||||
python3 .gitea/scripts/generate-k8s-wiki.py k8s/ Kubernetes-Services.md
|
.venv/bin/python .gitea/scripts/generate-k8s-wiki.py k8s/ Kubernetes-Services.md
|
||||||
|
|
||||||
if [ -f "Kubernetes-Services.md" ]; then
|
if [ -f "Kubernetes-Services.md" ]; then
|
||||||
echo "✅ Wiki content generated successfully"
|
echo "✅ Wiki content generated successfully"
|
||||||
|
|||||||
@@ -44,6 +44,7 @@ ArgoCD homelab project
|
|||||||
| **jellyfin** | [](https://ag.hexor.cy/applications/argocd/jellyfin) |
|
| **jellyfin** | [](https://ag.hexor.cy/applications/argocd/jellyfin) |
|
||||||
| **k8s-secrets** | [](https://ag.hexor.cy/applications/argocd/k8s-secrets) |
|
| **k8s-secrets** | [](https://ag.hexor.cy/applications/argocd/k8s-secrets) |
|
||||||
| **khm** | [](https://ag.hexor.cy/applications/argocd/khm) |
|
| **khm** | [](https://ag.hexor.cy/applications/argocd/khm) |
|
||||||
|
| **n8n** | [](https://ag.hexor.cy/applications/argocd/n8n) |
|
||||||
| **ollama** | [](https://ag.hexor.cy/applications/argocd/ollama) |
|
| **ollama** | [](https://ag.hexor.cy/applications/argocd/ollama) |
|
||||||
| **paperless** | [](https://ag.hexor.cy/applications/argocd/paperless) |
|
| **paperless** | [](https://ag.hexor.cy/applications/argocd/paperless) |
|
||||||
| **pasarguard** | [](https://ag.hexor.cy/applications/argocd/pasarguard) |
|
| **pasarguard** | [](https://ag.hexor.cy/applications/argocd/pasarguard) |
|
||||||
@@ -56,6 +57,7 @@ ArgoCD homelab project
|
|||||||
| **tg-bots** | [](https://ag.hexor.cy/applications/argocd/tg-bots) |
|
| **tg-bots** | [](https://ag.hexor.cy/applications/argocd/tg-bots) |
|
||||||
| **vaultwarden** | [](https://ag.hexor.cy/applications/argocd/vaultwarden) |
|
| **vaultwarden** | [](https://ag.hexor.cy/applications/argocd/vaultwarden) |
|
||||||
| **vpn** | [](https://ag.hexor.cy/applications/argocd/vpn) |
|
| **vpn** | [](https://ag.hexor.cy/applications/argocd/vpn) |
|
||||||
|
| **xandikos** | [](https://ag.hexor.cy/applications/argocd/xandikos) |
|
||||||
|
|
||||||
</td>
|
</td>
|
||||||
</tr>
|
</tr>
|
||||||
|
|||||||
@@ -77,8 +77,8 @@ spec:
|
|||||||
labels:
|
labels:
|
||||||
app: gitea-runner
|
app: gitea-runner
|
||||||
spec:
|
spec:
|
||||||
nodeSelector:
|
#nodeSelector:
|
||||||
kubernetes.io/hostname: home.homenet
|
# kubernetes.io/hostname: home.homenet
|
||||||
volumes:
|
volumes:
|
||||||
- name: docker-sock
|
- name: docker-sock
|
||||||
hostPath:
|
hostPath:
|
||||||
@@ -90,27 +90,30 @@ spec:
|
|||||||
affinity:
|
affinity:
|
||||||
nodeAffinity:
|
nodeAffinity:
|
||||||
preferredDuringSchedulingIgnoredDuringExecution:
|
preferredDuringSchedulingIgnoredDuringExecution:
|
||||||
- weight: 3
|
|
||||||
preference:
|
|
||||||
matchExpressions:
|
|
||||||
- key: kubernetes.io/hostname
|
|
||||||
operator: In
|
|
||||||
values:
|
|
||||||
- home.homenet
|
|
||||||
- weight: 1
|
- weight: 1
|
||||||
preference:
|
preference:
|
||||||
matchExpressions:
|
matchExpressions:
|
||||||
- key: kubernetes.io/hostname
|
- key: kubernetes.io/hostname
|
||||||
operator: In
|
operator: In
|
||||||
values:
|
values:
|
||||||
- master.tail2fe2d.ts.net
|
- home.homenet
|
||||||
- weight: 2
|
- weight: 2
|
||||||
preference:
|
preference:
|
||||||
matchExpressions:
|
matchExpressions:
|
||||||
- key: kubernetes.io/hostname
|
- key: kubernetes.io/hostname
|
||||||
operator: In
|
operator: In
|
||||||
values:
|
values:
|
||||||
- nas.homenet
|
- master.tail2fe2d.ts.net
|
||||||
|
- weight: 3
|
||||||
|
preference:
|
||||||
|
matchExpressions:
|
||||||
|
- key: kubernetes.io/hostname
|
||||||
|
operator: In
|
||||||
|
values:
|
||||||
|
- it.tail2fe2d.ts.net
|
||||||
|
- ch.tail2fe2d.ts.net
|
||||||
|
- us.tail2fe2d.ts.net
|
||||||
|
|
||||||
requiredDuringSchedulingIgnoredDuringExecution:
|
requiredDuringSchedulingIgnoredDuringExecution:
|
||||||
nodeSelectorTerms:
|
nodeSelectorTerms:
|
||||||
- matchExpressions:
|
- matchExpressions:
|
||||||
@@ -118,7 +121,9 @@ spec:
|
|||||||
operator: In
|
operator: In
|
||||||
values:
|
values:
|
||||||
- home.homenet
|
- home.homenet
|
||||||
- nas.homenet
|
- it.tail2fe2d.ts.net
|
||||||
|
- ch.tail2fe2d.ts.net
|
||||||
|
- us.tail2fe2d.ts.net
|
||||||
- master.tail2fe2d.ts.net
|
- master.tail2fe2d.ts.net
|
||||||
containers:
|
containers:
|
||||||
- name: gitea-runner
|
- name: gitea-runner
|
||||||
|
|||||||
@@ -74,19 +74,14 @@ spec:
|
|||||||
- nas.homenet
|
- nas.homenet
|
||||||
volumes:
|
volumes:
|
||||||
- name: upload-storage
|
- name: upload-storage
|
||||||
nfs:
|
persistentVolumeClaim:
|
||||||
server: nas.homenet
|
claimName: immich-upload-pvc
|
||||||
path: /mnt/storage/Storage/k8s/immich/library/
|
|
||||||
readOnly: false
|
|
||||||
- name: gphoto-storage
|
- name: gphoto-storage
|
||||||
nfs:
|
persistentVolumeClaim:
|
||||||
server: nas.homenet
|
claimName: immich-gphoto-pvc
|
||||||
path: /mnt/storage/Storage/k8s/immich/GPHOTO/
|
|
||||||
readOnly: false
|
|
||||||
- name: camera
|
- name: camera
|
||||||
nfs:
|
persistentVolumeClaim:
|
||||||
server: nas.homenet
|
claimName: immich-camera-pvc
|
||||||
path: /mnt/storage/Storage/Syncthing-repos/PhoneCamera/
|
|
||||||
readOnly: true
|
readOnly: true
|
||||||
- name: localtime
|
- name: localtime
|
||||||
hostPath:
|
hostPath:
|
||||||
|
|||||||
@@ -1,79 +1,52 @@
|
|||||||
---
|
---
|
||||||
apiVersion: v1
|
apiVersion: v1
|
||||||
kind: PersistentVolume
|
|
||||||
metadata:
|
|
||||||
name: immich-upload-pv
|
|
||||||
spec:
|
|
||||||
capacity:
|
|
||||||
storage: 500Gi
|
|
||||||
accessModes:
|
|
||||||
- ReadWriteOnce
|
|
||||||
hostPath:
|
|
||||||
path: /mnt/storage/Storage/k8s/immich/library
|
|
||||||
---
|
|
||||||
apiVersion: v1
|
|
||||||
kind: PersistentVolumeClaim
|
kind: PersistentVolumeClaim
|
||||||
metadata:
|
metadata:
|
||||||
name: immich-upload-pvc
|
name: immich-upload-pvc
|
||||||
namespace: immich
|
namespace: immich
|
||||||
spec:
|
spec:
|
||||||
storageClassName: ""
|
|
||||||
accessModes:
|
accessModes:
|
||||||
- ReadWriteOnce
|
- ReadWriteMany
|
||||||
volumeName: immich-upload-pv
|
storageClassName: nfs-csi
|
||||||
resources:
|
resources:
|
||||||
requests:
|
requests:
|
||||||
storage: 500Gi
|
storage: 500Gi
|
||||||
---
|
---
|
||||||
apiVersion: v1
|
apiVersion: v1
|
||||||
kind: PersistentVolume
|
|
||||||
metadata:
|
|
||||||
name: immich-gphoto-pv
|
|
||||||
spec:
|
|
||||||
capacity:
|
|
||||||
storage: 500Gi
|
|
||||||
accessModes:
|
|
||||||
- ReadWriteOnce
|
|
||||||
hostPath:
|
|
||||||
path: /mnt/storage/Storage/k8s/immich/GPHOTO
|
|
||||||
---
|
|
||||||
apiVersion: v1
|
|
||||||
kind: PersistentVolumeClaim
|
kind: PersistentVolumeClaim
|
||||||
metadata:
|
metadata:
|
||||||
name: immich-gphoto-pvc
|
name: immich-gphoto-pvc
|
||||||
namespace: immich
|
namespace: immich
|
||||||
spec:
|
spec:
|
||||||
storageClassName: ""
|
|
||||||
accessModes:
|
accessModes:
|
||||||
- ReadWriteOnce
|
- ReadWriteMany
|
||||||
volumeName: immich-gphoto-pv
|
storageClassName: nfs-csi
|
||||||
resources:
|
resources:
|
||||||
requests:
|
requests:
|
||||||
storage: 500Gi
|
storage: 500Gi
|
||||||
---
|
---
|
||||||
apiVersion: v1
|
apiVersion: v1
|
||||||
kind: PersistentVolume
|
|
||||||
metadata:
|
|
||||||
name: immich-db-pv
|
|
||||||
spec:
|
|
||||||
capacity:
|
|
||||||
storage: 10Gi
|
|
||||||
accessModes:
|
|
||||||
- ReadWriteOnce
|
|
||||||
hostPath:
|
|
||||||
path: /mnt/storage/Storage/k8s/immich/db
|
|
||||||
---
|
|
||||||
apiVersion: v1
|
|
||||||
kind: PersistentVolumeClaim
|
kind: PersistentVolumeClaim
|
||||||
metadata:
|
metadata:
|
||||||
name: immich-db-pvc
|
name: immich-db-pvc
|
||||||
namespace: immich
|
namespace: immich
|
||||||
spec:
|
spec:
|
||||||
storageClassName: ""
|
|
||||||
accessModes:
|
accessModes:
|
||||||
- ReadWriteOnce
|
- ReadWriteMany
|
||||||
volumeName: immich-db-pv
|
storageClassName: nfs-csi
|
||||||
resources:
|
resources:
|
||||||
requests:
|
requests:
|
||||||
storage: 10Gi
|
storage: 10Gi
|
||||||
|
---
|
||||||
|
apiVersion: v1
|
||||||
|
kind: PersistentVolumeClaim
|
||||||
|
metadata:
|
||||||
|
name: immich-camera-pvc
|
||||||
|
namespace: immich
|
||||||
|
spec:
|
||||||
|
accessModes:
|
||||||
|
- ReadOnlyMany
|
||||||
|
storageClassName: nfs-csi
|
||||||
|
resources:
|
||||||
|
requests:
|
||||||
|
storage: 100Gi
|
||||||
|
|||||||
@@ -19,7 +19,7 @@ spec:
|
|||||||
kubernetes.io/os: linux
|
kubernetes.io/os: linux
|
||||||
containers:
|
containers:
|
||||||
- name: secret-reader
|
- name: secret-reader
|
||||||
image: ultradesu/k8s-secrets:0.1.1
|
image: ultradesu/k8s-secrets:0.2.1
|
||||||
imagePullPolicy: Always
|
imagePullPolicy: Always
|
||||||
args:
|
args:
|
||||||
- "--secrets"
|
- "--secrets"
|
||||||
@@ -28,6 +28,7 @@ spec:
|
|||||||
- "k8s-secret"
|
- "k8s-secret"
|
||||||
- "--port"
|
- "--port"
|
||||||
- "3000"
|
- "3000"
|
||||||
|
- "--webhook"
|
||||||
ports:
|
ports:
|
||||||
- containerPort: 3000
|
- containerPort: 3000
|
||||||
name: http
|
name: http
|
||||||
|
|||||||
21
k8s/apps/n8n/app.yaml
Normal file
21
k8s/apps/n8n/app.yaml
Normal file
@@ -0,0 +1,21 @@
|
|||||||
|
apiVersion: argoproj.io/v1alpha1
|
||||||
|
kind: Application
|
||||||
|
metadata:
|
||||||
|
name: n8n
|
||||||
|
namespace: argocd
|
||||||
|
spec:
|
||||||
|
project: apps
|
||||||
|
destination:
|
||||||
|
namespace: n8n
|
||||||
|
server: https://kubernetes.default.svc
|
||||||
|
source:
|
||||||
|
repoURL: ssh://git@gt.hexor.cy:30022/ab/homelab.git
|
||||||
|
targetRevision: HEAD
|
||||||
|
path: k8s/apps/n8n
|
||||||
|
syncPolicy:
|
||||||
|
automated:
|
||||||
|
selfHeal: true
|
||||||
|
prune: true
|
||||||
|
syncOptions:
|
||||||
|
- CreateNamespace=true
|
||||||
|
|
||||||
39
k8s/apps/n8n/external-secrets.yaml
Normal file
39
k8s/apps/n8n/external-secrets.yaml
Normal file
@@ -0,0 +1,39 @@
|
|||||||
|
---
|
||||||
|
apiVersion: external-secrets.io/v1
|
||||||
|
kind: ExternalSecret
|
||||||
|
metadata:
|
||||||
|
name: credentials
|
||||||
|
spec:
|
||||||
|
target:
|
||||||
|
name: credentials
|
||||||
|
deletionPolicy: Delete
|
||||||
|
template:
|
||||||
|
type: Opaque
|
||||||
|
data:
|
||||||
|
postgres-password: |-
|
||||||
|
{{ .psql }}
|
||||||
|
N8N_ENCRYPTION_KEY: |-
|
||||||
|
{{ .enc-pass }}
|
||||||
|
data:
|
||||||
|
- secretKey: psql
|
||||||
|
sourceRef:
|
||||||
|
storeRef:
|
||||||
|
name: vaultwarden-login
|
||||||
|
kind: ClusterSecretStore
|
||||||
|
remoteRef:
|
||||||
|
conversionStrategy: Default
|
||||||
|
decodingStrategy: None
|
||||||
|
metadataPolicy: None
|
||||||
|
key: 2a9deb39-ef22-433e-a1be-df1555625e22
|
||||||
|
property: fields[13].value
|
||||||
|
- secretKey: enc-pass
|
||||||
|
sourceRef:
|
||||||
|
storeRef:
|
||||||
|
name: vaultwarden-login
|
||||||
|
kind: ClusterSecretStore
|
||||||
|
remoteRef:
|
||||||
|
conversionStrategy: Default
|
||||||
|
decodingStrategy: None
|
||||||
|
metadataPolicy: None
|
||||||
|
key: 18c92d73-9637-4419-8642-7f7b308460cb
|
||||||
|
property: fields[0].value
|
||||||
22
k8s/apps/n8n/kustomization.yaml
Normal file
22
k8s/apps/n8n/kustomization.yaml
Normal file
@@ -0,0 +1,22 @@
|
|||||||
|
apiVersion: kustomize.config.k8s.io/v1beta1
|
||||||
|
kind: Kustomization
|
||||||
|
|
||||||
|
resources:
|
||||||
|
- external-secrets.yaml
|
||||||
|
- storage.yaml
|
||||||
|
|
||||||
|
helmCharts:
|
||||||
|
- name: n8n
|
||||||
|
repo: https://community-charts.github.io/helm-charts
|
||||||
|
version: 1.16.28
|
||||||
|
releaseName: n8n
|
||||||
|
namespace: n8n
|
||||||
|
valuesFile: values-n8n.yaml
|
||||||
|
includeCRDs: true
|
||||||
|
- name: searxng
|
||||||
|
repo: https://unknowniq.github.io/helm-charts/
|
||||||
|
version: 0.1.3
|
||||||
|
releaseName: searxng
|
||||||
|
namespace: n8n
|
||||||
|
valuesFile: values-searxng.yaml
|
||||||
|
includeCRDs: true
|
||||||
12
k8s/apps/n8n/storage.yaml
Normal file
12
k8s/apps/n8n/storage.yaml
Normal file
@@ -0,0 +1,12 @@
|
|||||||
|
---
|
||||||
|
apiVersion: v1
|
||||||
|
kind: PersistentVolumeClaim
|
||||||
|
metadata:
|
||||||
|
name: n8n-home
|
||||||
|
spec:
|
||||||
|
accessModes:
|
||||||
|
- ReadWriteMany
|
||||||
|
storageClassName: nfs-csi
|
||||||
|
resources:
|
||||||
|
requests:
|
||||||
|
storage: 10Gi
|
||||||
59
k8s/apps/n8n/values-n8n.yaml
Normal file
59
k8s/apps/n8n/values-n8n.yaml
Normal file
@@ -0,0 +1,59 @@
|
|||||||
|
nodeSelector:
|
||||||
|
kubernetes.io/hostname: master.tail2fe2d.ts.net
|
||||||
|
|
||||||
|
db:
|
||||||
|
type: postgresdb
|
||||||
|
|
||||||
|
main:
|
||||||
|
resources:
|
||||||
|
requests:
|
||||||
|
cpu: 100m
|
||||||
|
memory: 128Mi
|
||||||
|
limits:
|
||||||
|
cpu: 512m
|
||||||
|
memory: 512Mi
|
||||||
|
persistence:
|
||||||
|
enabled: true
|
||||||
|
existingClaim: n8n-home
|
||||||
|
mountPath: /home/node/.n8n
|
||||||
|
|
||||||
|
worker:
|
||||||
|
mode: queue
|
||||||
|
persistence:
|
||||||
|
enabled: true
|
||||||
|
existingClaim: n8n-home
|
||||||
|
mountPath: /home/node/.n8n
|
||||||
|
|
||||||
|
webhook:
|
||||||
|
url: https://n8n.hexor.cy
|
||||||
|
persistence:
|
||||||
|
enabled: true
|
||||||
|
existingClaim: n8n-home
|
||||||
|
mountPath: /home/node/.n8n
|
||||||
|
|
||||||
|
redis:
|
||||||
|
enabled: true
|
||||||
|
|
||||||
|
externalPostgresql:
|
||||||
|
existingSecret: postgres-password
|
||||||
|
host: "psql.psql.svc"
|
||||||
|
username: "n8n"
|
||||||
|
database: "n8n"
|
||||||
|
|
||||||
|
ingress:
|
||||||
|
enabled: true
|
||||||
|
className: traefik
|
||||||
|
annotations:
|
||||||
|
cert-manager.io/cluster-issuer: letsencrypt
|
||||||
|
traefik.ingress.kubernetes.io/router.middlewares: kube-system-https-redirect@kubernetescrd
|
||||||
|
|
||||||
|
hosts:
|
||||||
|
- host: n8n.hexor.cy
|
||||||
|
paths:
|
||||||
|
- path: /
|
||||||
|
pathType: Prefix
|
||||||
|
tls:
|
||||||
|
- secretName: n8n-tls
|
||||||
|
hosts:
|
||||||
|
- '*.hexor.cy'
|
||||||
|
|
||||||
8
k8s/apps/n8n/values-searxng.yaml
Normal file
8
k8s/apps/n8n/values-searxng.yaml
Normal file
@@ -0,0 +1,8 @@
|
|||||||
|
config:
|
||||||
|
general:
|
||||||
|
instance_name: "HexorSearXNG"
|
||||||
|
|
||||||
|
valkey:
|
||||||
|
enabled: true
|
||||||
|
nodeSelector:
|
||||||
|
kubernetes.io/hostname: master.tail2fe2d.ts.net
|
||||||
@@ -4,6 +4,7 @@ kind: Kustomization
|
|||||||
resources:
|
resources:
|
||||||
- app.yaml
|
- app.yaml
|
||||||
- external-secrets.yaml
|
- external-secrets.yaml
|
||||||
|
- paperless-ai.yaml
|
||||||
|
|
||||||
helmCharts:
|
helmCharts:
|
||||||
- name: paperless-ngx
|
- name: paperless-ngx
|
||||||
|
|||||||
101
k8s/apps/paperless/paperless-ai.yaml
Normal file
101
k8s/apps/paperless/paperless-ai.yaml
Normal file
@@ -0,0 +1,101 @@
|
|||||||
|
---
|
||||||
|
apiVersion: apps/v1
|
||||||
|
kind: Deployment
|
||||||
|
metadata:
|
||||||
|
name: paperless-ai
|
||||||
|
labels:
|
||||||
|
app: paperless-ai
|
||||||
|
spec:
|
||||||
|
replicas: 1
|
||||||
|
selector:
|
||||||
|
matchLabels:
|
||||||
|
app: paperless-ai
|
||||||
|
template:
|
||||||
|
metadata:
|
||||||
|
labels:
|
||||||
|
app: paperless-ai
|
||||||
|
spec:
|
||||||
|
nodeSelector:
|
||||||
|
kubernetes.io/hostname: nas.homenet
|
||||||
|
containers:
|
||||||
|
- name: paperless-ai
|
||||||
|
image: clusterzx/paperless-ai:latest
|
||||||
|
imagePullPolicy: Always
|
||||||
|
ports:
|
||||||
|
- containerPort: 3000
|
||||||
|
name: http
|
||||||
|
env:
|
||||||
|
- name: NODE_ENV
|
||||||
|
value: production
|
||||||
|
- name: PAPERLESS_AI_PORT
|
||||||
|
value: "3000"
|
||||||
|
resources:
|
||||||
|
requests:
|
||||||
|
memory: 512Mi
|
||||||
|
cpu: 500m
|
||||||
|
limits:
|
||||||
|
memory: 1024Mi
|
||||||
|
cpu: 2000m
|
||||||
|
#livenessProbe:
|
||||||
|
# httpGet:
|
||||||
|
# path: /
|
||||||
|
# port: 8000
|
||||||
|
# initialDelaySeconds: 30
|
||||||
|
# periodSeconds: 10
|
||||||
|
#readinessProbe:
|
||||||
|
# httpGet:
|
||||||
|
# path: /
|
||||||
|
# port: 8000
|
||||||
|
# initialDelaySeconds: 5
|
||||||
|
# periodSeconds: 5
|
||||||
|
volumeMounts:
|
||||||
|
- name: data
|
||||||
|
mountPath: /app/data
|
||||||
|
volumes:
|
||||||
|
- name: data
|
||||||
|
hostPath:
|
||||||
|
path: /mnt/storage/Storage/k8s/paperless/ai-data
|
||||||
|
type: DirectoryOrCreate
|
||||||
|
---
|
||||||
|
apiVersion: v1
|
||||||
|
kind: Service
|
||||||
|
metadata:
|
||||||
|
name: paperless-ai
|
||||||
|
namespace: paperless
|
||||||
|
labels:
|
||||||
|
app: paperless-ai
|
||||||
|
spec:
|
||||||
|
type: ClusterIP
|
||||||
|
ports:
|
||||||
|
- port: 3000
|
||||||
|
targetPort: 3000
|
||||||
|
protocol: TCP
|
||||||
|
name: http
|
||||||
|
selector:
|
||||||
|
app: paperless-ai
|
||||||
|
---
|
||||||
|
apiVersion: networking.k8s.io/v1
|
||||||
|
kind: Ingress
|
||||||
|
metadata:
|
||||||
|
name: paperless-ai-ingress
|
||||||
|
annotations:
|
||||||
|
ingressClassName: traefik
|
||||||
|
cert-manager.io/cluster-issuer: letsencrypt
|
||||||
|
traefik.ingress.kubernetes.io/router.middlewares: kube-system-https-redirect@kubernetescrd
|
||||||
|
acme.cert-manager.io/http01-edit-in-place: "true"
|
||||||
|
spec:
|
||||||
|
rules:
|
||||||
|
- host: ai-docs.hexor.cy
|
||||||
|
http:
|
||||||
|
paths:
|
||||||
|
- path: /
|
||||||
|
pathType: Prefix
|
||||||
|
backend:
|
||||||
|
service:
|
||||||
|
name: paperless-ai
|
||||||
|
port:
|
||||||
|
number: 3000
|
||||||
|
tls:
|
||||||
|
- secretName: docs-tls
|
||||||
|
hosts:
|
||||||
|
- '*.hexor.cy'
|
||||||
@@ -1,5 +1,5 @@
|
|||||||
image:
|
image:
|
||||||
tag: 2.19.3
|
tag: 2.20.3
|
||||||
resources:
|
resources:
|
||||||
requests:
|
requests:
|
||||||
memory: "1Gi"
|
memory: "1Gi"
|
||||||
|
|||||||
@@ -1,212 +0,0 @@
|
|||||||
---
|
|
||||||
apiVersion: v1
|
|
||||||
kind: ConfigMap
|
|
||||||
metadata:
|
|
||||||
name: pasarguard-scripts-ingress
|
|
||||||
labels:
|
|
||||||
app: pasarguard-node-ingress
|
|
||||||
data:
|
|
||||||
init-uuid-ingress.sh: |
|
|
||||||
#!/bin/bash
|
|
||||||
set -e
|
|
||||||
echo "Started"
|
|
||||||
# NODE_NAME is already set via environment variable
|
|
||||||
NAMESPACE=$(cat /var/run/secrets/kubernetes.io/serviceaccount/namespace)
|
|
||||||
|
|
||||||
# Get DNS name from node label xray-public-address
|
|
||||||
DNS_NAME=$(kubectl get node "${NODE_NAME}" -o jsonpath='{.metadata.labels.xray-public-address}')
|
|
||||||
|
|
||||||
if [ -z "${DNS_NAME}" ]; then
|
|
||||||
echo "ERROR: Node ${NODE_NAME} does not have label 'xray-public-address'"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
echo "Node: ${NODE_NAME}"
|
|
||||||
echo "DNS Name from label: ${DNS_NAME}"
|
|
||||||
|
|
||||||
# Use DNS name for ConfigMap name to ensure uniqueness
|
|
||||||
CONFIGMAP_NAME="node-uuid-ingress-${DNS_NAME//./-}"
|
|
||||||
|
|
||||||
echo "Checking ConfigMap: ${CONFIGMAP_NAME}"
|
|
||||||
|
|
||||||
# Check if ConfigMap exists and get UUID
|
|
||||||
if kubectl get configmap "${CONFIGMAP_NAME}" -n "${NAMESPACE}" &>/dev/null; then
|
|
||||||
echo "ConfigMap exists, reading UUID..."
|
|
||||||
API_KEY=$(kubectl get configmap "${CONFIGMAP_NAME}" -n "${NAMESPACE}" -o jsonpath='{.data.API_KEY}')
|
|
||||||
|
|
||||||
if [ -z "${API_KEY}" ]; then
|
|
||||||
echo "UUID not found in ConfigMap, generating new one..."
|
|
||||||
API_KEY=$(cat /proc/sys/kernel/random/uuid)
|
|
||||||
kubectl patch configmap "${CONFIGMAP_NAME}" -n "${NAMESPACE}" --type merge -p "{\"data\":{\"API_KEY\":\"${API_KEY}\"}}"
|
|
||||||
else
|
|
||||||
echo "Using existing UUID from ConfigMap"
|
|
||||||
fi
|
|
||||||
else
|
|
||||||
echo "ConfigMap does not exist, creating new one..."
|
|
||||||
API_KEY=$(cat /proc/sys/kernel/random/uuid)
|
|
||||||
kubectl create configmap "${CONFIGMAP_NAME}" -n "${NAMESPACE}" \
|
|
||||||
--from-literal=API_KEY="${API_KEY}" \
|
|
||||||
--from-literal=NODE_NAME="${NODE_NAME}"
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Save UUID and node info to shared volume for the main container
|
|
||||||
echo -n "${API_KEY}" > /shared/api-key
|
|
||||||
echo -n "${NODE_NAME}" > /shared/node-name
|
|
||||||
echo -n "${CONFIGMAP_NAME}" > /shared/configmap-name
|
|
||||||
echo "UUID initialized: ${API_KEY}"
|
|
||||||
echo "Node name: ${NODE_NAME}"
|
|
||||||
echo "ConfigMap: ${CONFIGMAP_NAME}"
|
|
||||||
|
|
||||||
# Create Certificate for this node using DNS name from label
|
|
||||||
CERT_NAME="pasarguard-node-ingress-${DNS_NAME//./-}"
|
|
||||||
|
|
||||||
echo "Creating Certificate: ${CERT_NAME} for ${DNS_NAME}"
|
|
||||||
|
|
||||||
# Check if Certificate already exists
|
|
||||||
if ! kubectl get certificate "${CERT_NAME}" -n "${NAMESPACE}" &>/dev/null; then
|
|
||||||
echo "Certificate does not exist, creating..."
|
|
||||||
cat <<EOF | kubectl apply -f -
|
|
||||||
apiVersion: cert-manager.io/v1
|
|
||||||
kind: Certificate
|
|
||||||
metadata:
|
|
||||||
name: ${CERT_NAME}
|
|
||||||
namespace: ${NAMESPACE}
|
|
||||||
spec:
|
|
||||||
secretName: ${CERT_NAME}-tls
|
|
||||||
issuerRef:
|
|
||||||
name: letsencrypt
|
|
||||||
kind: ClusterIssuer
|
|
||||||
dnsNames:
|
|
||||||
- ${DNS_NAME}
|
|
||||||
EOF
|
|
||||||
else
|
|
||||||
echo "Certificate already exists"
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Wait for certificate to be ready
|
|
||||||
|
|
||||||
echo "Waiting for certificate to be ready..."
|
|
||||||
for i in {1..600}; do
|
|
||||||
if kubectl get secret "${CERT_NAME}-tls" -n "${NAMESPACE}" &>/dev/null; then
|
|
||||||
echo "Certificate secret is ready!"
|
|
||||||
break
|
|
||||||
fi
|
|
||||||
echo "Waiting for certificate... ($i/600)"
|
|
||||||
sleep 1
|
|
||||||
done
|
|
||||||
|
|
||||||
if ! kubectl get secret "${CERT_NAME}-tls" -n "${NAMESPACE}" &>/dev/null; then
|
|
||||||
echo "WARNING: Certificate secret not ready after 600 seconds"
|
|
||||||
else
|
|
||||||
# Extract certificate and key from secret to shared volume
|
|
||||||
echo "Extracting certificate and key..."
|
|
||||||
kubectl get secret "${CERT_NAME}-tls" -n "${NAMESPACE}" -o jsonpath='{.data.tls\.crt}' | base64 -d > /shared/tls.crt
|
|
||||||
kubectl get secret "${CERT_NAME}-tls" -n "${NAMESPACE}" -o jsonpath='{.data.tls\.key}' | base64 -d > /shared/tls.key
|
|
||||||
echo "Certificate and key extracted successfully."
|
|
||||||
cat /shared/tls.crt
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Create ClusterIP Service for this node (pod selector based)
|
|
||||||
NODE_SHORT_NAME="${NODE_NAME%%.*}"
|
|
||||||
SERVICE_NAME="${NODE_SHORT_NAME}-ingress"
|
|
||||||
|
|
||||||
echo "Creating Service: ${SERVICE_NAME} for node ${NODE_NAME} (short: ${NODE_SHORT_NAME})"
|
|
||||||
|
|
||||||
# Create Service with pod selector including node name
|
|
||||||
cat <<EOF | kubectl apply -f -
|
|
||||||
apiVersion: v1
|
|
||||||
kind: Service
|
|
||||||
metadata:
|
|
||||||
name: ${SERVICE_NAME}
|
|
||||||
namespace: ${NAMESPACE}
|
|
||||||
labels:
|
|
||||||
app: pasarguard-node-ingress
|
|
||||||
node: ${NODE_NAME}
|
|
||||||
spec:
|
|
||||||
type: ClusterIP
|
|
||||||
selector:
|
|
||||||
app: pasarguard-node-ingress
|
|
||||||
node-name: ${NODE_SHORT_NAME}
|
|
||||||
ports:
|
|
||||||
- name: proxy
|
|
||||||
port: 443
|
|
||||||
protocol: TCP
|
|
||||||
targetPort: 443
|
|
||||||
- name: api
|
|
||||||
port: 62050
|
|
||||||
protocol: TCP
|
|
||||||
targetPort: 62050
|
|
||||||
EOF
|
|
||||||
|
|
||||||
echo "Service created: ${SERVICE_NAME}.${NAMESPACE}.svc.cluster.local"
|
|
||||||
|
|
||||||
# Create IngressRouteTCP for this DNS name with TLS passthrough
|
|
||||||
INGRESS_NAME="pasarguard-tcp-${DNS_NAME//./-}"
|
|
||||||
|
|
||||||
echo "Creating IngressRouteTCP: ${INGRESS_NAME} for ${DNS_NAME}"
|
|
||||||
|
|
||||||
cat <<EOF | kubectl apply -f -
|
|
||||||
apiVersion: traefik.io/v1alpha1
|
|
||||||
kind: IngressRouteTCP
|
|
||||||
metadata:
|
|
||||||
name: ${INGRESS_NAME}
|
|
||||||
namespace: ${NAMESPACE}
|
|
||||||
labels:
|
|
||||||
app: pasarguard-node-ingress
|
|
||||||
node: ${NODE_NAME}
|
|
||||||
spec:
|
|
||||||
entryPoints:
|
|
||||||
- websecure
|
|
||||||
routes:
|
|
||||||
- match: HostSNI(\`${DNS_NAME}\`)
|
|
||||||
services:
|
|
||||||
- name: ${SERVICE_NAME}
|
|
||||||
port: 443
|
|
||||||
tls:
|
|
||||||
passthrough: true
|
|
||||||
EOF
|
|
||||||
|
|
||||||
echo "IngressRouteTCP created: ${INGRESS_NAME}"
|
|
||||||
echo "Traffic to ${DNS_NAME}:443 will be routed to ${SERVICE_NAME}:443"
|
|
||||||
|
|
||||||
# Create second IngressRouteTCP for API port 62051
|
|
||||||
INGRESS_API_NAME="pasarguard-api-${DNS_NAME//./-}"
|
|
||||||
|
|
||||||
echo "Creating IngressRouteTCP for API: ${INGRESS_API_NAME} for ${DNS_NAME}:62051"
|
|
||||||
|
|
||||||
cat <<EOF | kubectl apply -f -
|
|
||||||
apiVersion: traefik.io/v1alpha1
|
|
||||||
kind: IngressRouteTCP
|
|
||||||
metadata:
|
|
||||||
name: ${INGRESS_API_NAME}
|
|
||||||
namespace: ${NAMESPACE}
|
|
||||||
labels:
|
|
||||||
app: pasarguard-node-ingress
|
|
||||||
node: ${NODE_NAME}
|
|
||||||
spec:
|
|
||||||
entryPoints:
|
|
||||||
- pasarguard-api
|
|
||||||
routes:
|
|
||||||
- match: HostSNI(\`${DNS_NAME}\`)
|
|
||||||
services:
|
|
||||||
- name: ${SERVICE_NAME}
|
|
||||||
port: 62050
|
|
||||||
tls:
|
|
||||||
passthrough: true
|
|
||||||
EOF
|
|
||||||
|
|
||||||
echo "IngressRouteTCP API created: ${INGRESS_API_NAME}"
|
|
||||||
echo "Traffic to ${DNS_NAME}:62051 will be routed to ${SERVICE_NAME}:62050"
|
|
||||||
|
|
||||||
pasarguard-start.sh: |
|
|
||||||
#!/bin/sh
|
|
||||||
# Read API_KEY from shared volume created by init container
|
|
||||||
if [ -f /shared/api-key ]; then
|
|
||||||
export API_KEY=$(cat /shared/api-key)
|
|
||||||
echo "Loaded API_KEY from shared volume"
|
|
||||||
else
|
|
||||||
echo "WARNING: API_KEY file not found, using default"
|
|
||||||
fi
|
|
||||||
|
|
||||||
cd /app
|
|
||||||
exec ./main
|
|
||||||
@@ -1,211 +0,0 @@
|
|||||||
---
|
|
||||||
apiVersion: v1
|
|
||||||
kind: ServiceAccount
|
|
||||||
metadata:
|
|
||||||
name: pasarguard-node-ingress
|
|
||||||
labels:
|
|
||||||
app: pasarguard-node-ingress
|
|
||||||
---
|
|
||||||
apiVersion: rbac.authorization.k8s.io/v1
|
|
||||||
kind: Role
|
|
||||||
metadata:
|
|
||||||
name: pasarguard-node-ingress-configmap
|
|
||||||
labels:
|
|
||||||
app: pasarguard-node-ingress
|
|
||||||
rules:
|
|
||||||
- apiGroups: [""]
|
|
||||||
resources: ["configmaps"]
|
|
||||||
verbs: ["get", "list", "create", "update", "patch"]
|
|
||||||
- apiGroups: ["cert-manager.io"]
|
|
||||||
resources: ["certificates"]
|
|
||||||
verbs: ["get", "list", "create", "update", "patch", "delete"]
|
|
||||||
- apiGroups: [""]
|
|
||||||
resources: ["secrets"]
|
|
||||||
verbs: ["get", "list"]
|
|
||||||
- apiGroups: [""]
|
|
||||||
resources: ["services", "endpoints"]
|
|
||||||
verbs: ["get", "list", "create", "update", "patch", "delete"]
|
|
||||||
- apiGroups: ["traefik.io", "traefik.containo.us"]
|
|
||||||
resources: ["ingressroutetcps"]
|
|
||||||
verbs: ["get", "list", "create", "update", "patch", "delete"]
|
|
||||||
- apiGroups: [""]
|
|
||||||
resources: ["pods"]
|
|
||||||
verbs: ["get", "list", "patch", "update"]
|
|
||||||
---
|
|
||||||
apiVersion: rbac.authorization.k8s.io/v1
|
|
||||||
kind: RoleBinding
|
|
||||||
metadata:
|
|
||||||
name: pasarguard-node-ingress-configmap
|
|
||||||
labels:
|
|
||||||
app: pasarguard-node-ingress
|
|
||||||
roleRef:
|
|
||||||
apiGroup: rbac.authorization.k8s.io
|
|
||||||
kind: Role
|
|
||||||
name: pasarguard-node-ingress-configmap
|
|
||||||
subjects:
|
|
||||||
- kind: ServiceAccount
|
|
||||||
name: pasarguard-node-ingress
|
|
||||||
---
|
|
||||||
apiVersion: rbac.authorization.k8s.io/v1
|
|
||||||
kind: ClusterRole
|
|
||||||
metadata:
|
|
||||||
name: pasarguard-node-ingress-reader
|
|
||||||
labels:
|
|
||||||
app: pasarguard-node-ingress
|
|
||||||
rules:
|
|
||||||
- apiGroups: [""]
|
|
||||||
resources: ["nodes"]
|
|
||||||
verbs: ["get", "list"]
|
|
||||||
---
|
|
||||||
apiVersion: rbac.authorization.k8s.io/v1
|
|
||||||
kind: ClusterRoleBinding
|
|
||||||
metadata:
|
|
||||||
name: pasarguard-node-ingress-reader
|
|
||||||
labels:
|
|
||||||
app: pasarguard-node-ingress
|
|
||||||
roleRef:
|
|
||||||
apiGroup: rbac.authorization.k8s.io
|
|
||||||
kind: ClusterRole
|
|
||||||
name: pasarguard-node-ingress-reader
|
|
||||||
subjects:
|
|
||||||
- kind: ServiceAccount
|
|
||||||
name: pasarguard-node-ingress
|
|
||||||
namespace: pasarguard
|
|
||||||
---
|
|
||||||
apiVersion: apps/v1
|
|
||||||
kind: DaemonSet
|
|
||||||
metadata:
|
|
||||||
name: pasarguard-node-ingress
|
|
||||||
labels:
|
|
||||||
app: pasarguard-node-ingress
|
|
||||||
spec:
|
|
||||||
selector:
|
|
||||||
matchLabels:
|
|
||||||
app: pasarguard-node-ingress
|
|
||||||
revisionHistoryLimit: 3
|
|
||||||
updateStrategy:
|
|
||||||
type: RollingUpdate
|
|
||||||
template:
|
|
||||||
metadata:
|
|
||||||
labels:
|
|
||||||
app: pasarguard-node-ingress
|
|
||||||
spec:
|
|
||||||
serviceAccountName: pasarguard-node-ingress
|
|
||||||
affinity:
|
|
||||||
nodeAffinity:
|
|
||||||
requiredDuringSchedulingIgnoredDuringExecution:
|
|
||||||
nodeSelectorTerms:
|
|
||||||
- matchExpressions:
|
|
||||||
- key: xray-public-address
|
|
||||||
operator: Exists
|
|
||||||
initContainers:
|
|
||||||
- name: label-pod
|
|
||||||
image: bitnami/kubectl:latest
|
|
||||||
env:
|
|
||||||
- name: POD_NAME
|
|
||||||
valueFrom:
|
|
||||||
fieldRef:
|
|
||||||
fieldPath: metadata.name
|
|
||||||
- name: POD_NAMESPACE
|
|
||||||
valueFrom:
|
|
||||||
fieldRef:
|
|
||||||
fieldPath: metadata.namespace
|
|
||||||
- name: NODE_NAME
|
|
||||||
valueFrom:
|
|
||||||
fieldRef:
|
|
||||||
fieldPath: spec.nodeName
|
|
||||||
command:
|
|
||||||
- /bin/bash
|
|
||||||
- -c
|
|
||||||
- |
|
|
||||||
# Add node label to pod
|
|
||||||
NODE_SHORT=$(echo ${NODE_NAME} | cut -d. -f1)
|
|
||||||
kubectl label pod ${POD_NAME} -n ${POD_NAMESPACE} node-name=${NODE_SHORT} --overwrite
|
|
||||||
- name: init-uuid
|
|
||||||
image: bitnami/kubectl:latest
|
|
||||||
env:
|
|
||||||
- name: GODEBUG
|
|
||||||
value: "x509sha1=1"
|
|
||||||
- name: NODE_NAME
|
|
||||||
valueFrom:
|
|
||||||
fieldRef:
|
|
||||||
fieldPath: spec.nodeName
|
|
||||||
command:
|
|
||||||
- /bin/bash
|
|
||||||
- /scripts/init-uuid-ingress.sh
|
|
||||||
volumeMounts:
|
|
||||||
- name: shared-data
|
|
||||||
mountPath: /shared
|
|
||||||
- name: scripts
|
|
||||||
mountPath: /scripts
|
|
||||||
containers:
|
|
||||||
- name: pasarguard-node
|
|
||||||
image: 'pasarguard/node:v0.1.3'
|
|
||||||
imagePullPolicy: Always
|
|
||||||
command:
|
|
||||||
- /bin/sh
|
|
||||||
- /scripts/pasarguard-start.sh
|
|
||||||
ports:
|
|
||||||
- name: api
|
|
||||||
containerPort: 62050
|
|
||||||
protocol: TCP
|
|
||||||
- name: proxy
|
|
||||||
containerPort: 443
|
|
||||||
protocol: TCP
|
|
||||||
env:
|
|
||||||
- name: NODE_NAME
|
|
||||||
valueFrom:
|
|
||||||
fieldRef:
|
|
||||||
fieldPath: spec.nodeName
|
|
||||||
- name: NODE_HOST
|
|
||||||
value: "0.0.0.0"
|
|
||||||
- name: SERVICE_PORT
|
|
||||||
value: "62050"
|
|
||||||
- name: SERVICE_PROTOCOL
|
|
||||||
value: "grpc"
|
|
||||||
- name: DEBUG
|
|
||||||
value: "true"
|
|
||||||
- name: SSL_CERT_FILE
|
|
||||||
value: "/shared/tls.crt"
|
|
||||||
- name: SSL_KEY_FILE
|
|
||||||
value: "/shared/tls.key"
|
|
||||||
- name: XRAY_EXECUTABLE_PATH
|
|
||||||
value: "/usr/local/bin/xray"
|
|
||||||
- name: XRAY_ASSETS_PATH
|
|
||||||
value: "/usr/local/share/xray"
|
|
||||||
- name: API_KEY
|
|
||||||
value: "change-this-to-a-secure-uuid"
|
|
||||||
livenessProbe:
|
|
||||||
tcpSocket:
|
|
||||||
port: 62050
|
|
||||||
initialDelaySeconds: 30
|
|
||||||
periodSeconds: 10
|
|
||||||
timeoutSeconds: 5
|
|
||||||
failureThreshold: 3
|
|
||||||
readinessProbe:
|
|
||||||
tcpSocket:
|
|
||||||
port: 62050
|
|
||||||
initialDelaySeconds: 10
|
|
||||||
periodSeconds: 5
|
|
||||||
timeoutSeconds: 3
|
|
||||||
failureThreshold: 3
|
|
||||||
resources:
|
|
||||||
requests:
|
|
||||||
memory: "128Mi"
|
|
||||||
cpu: "100m"
|
|
||||||
limits:
|
|
||||||
memory: "512Mi"
|
|
||||||
cpu: "750m"
|
|
||||||
volumeMounts:
|
|
||||||
- name: shared-data
|
|
||||||
mountPath: /shared
|
|
||||||
readOnly: false
|
|
||||||
- name: scripts
|
|
||||||
mountPath: /scripts
|
|
||||||
volumes:
|
|
||||||
- name: shared-data
|
|
||||||
emptyDir: {}
|
|
||||||
- name: scripts
|
|
||||||
configMap:
|
|
||||||
name: pasarguard-scripts-ingress
|
|
||||||
defaultMode: 0755
|
|
||||||
@@ -113,7 +113,7 @@ spec:
|
|||||||
mountPath: /scripts
|
mountPath: /scripts
|
||||||
containers:
|
containers:
|
||||||
- name: pasarguard-node
|
- name: pasarguard-node
|
||||||
image: 'pasarguard/node:v0.1.3'
|
image: 'pasarguard/node:v0.2.1'
|
||||||
imagePullPolicy: Always
|
imagePullPolicy: Always
|
||||||
command:
|
command:
|
||||||
- /bin/sh
|
- /bin/sh
|
||||||
@@ -162,10 +162,10 @@ spec:
|
|||||||
resources:
|
resources:
|
||||||
requests:
|
requests:
|
||||||
memory: "128Mi"
|
memory: "128Mi"
|
||||||
cpu: "100m"
|
#cpu: "500m"
|
||||||
limits:
|
limits:
|
||||||
memory: "512Mi"
|
memory: "512Mi"
|
||||||
cpu: "750m"
|
#cpu: "1200m"
|
||||||
volumeMounts:
|
volumeMounts:
|
||||||
- name: shared-data
|
- name: shared-data
|
||||||
mountPath: /shared
|
mountPath: /shared
|
||||||
@@ -205,7 +205,7 @@ spec:
|
|||||||
cpu: "50m"
|
cpu: "50m"
|
||||||
limits:
|
limits:
|
||||||
memory: "128Mi"
|
memory: "128Mi"
|
||||||
cpu: "150m"
|
cpu: "500m"
|
||||||
volumeMounts:
|
volumeMounts:
|
||||||
- name: shared-data
|
- name: shared-data
|
||||||
mountPath: /shared
|
mountPath: /shared
|
||||||
|
|||||||
@@ -9,6 +9,3 @@ resources:
|
|||||||
- ./certificate.yaml
|
- ./certificate.yaml
|
||||||
- ./configmap-scripts.yaml
|
- ./configmap-scripts.yaml
|
||||||
- ./servicemonitor.yaml
|
- ./servicemonitor.yaml
|
||||||
- ./configmap-scripts-ingress.yaml
|
|
||||||
# - ./daemonset-ingress.yaml
|
|
||||||
# - ./traefik-pasarguard-entrypoint.yaml
|
|
||||||
|
|||||||
@@ -1,66 +0,0 @@
|
|||||||
---
|
|
||||||
apiVersion: apps/v1
|
|
||||||
kind: Deployment
|
|
||||||
metadata:
|
|
||||||
name: traefik
|
|
||||||
namespace: kube-system
|
|
||||||
spec:
|
|
||||||
template:
|
|
||||||
spec:
|
|
||||||
containers:
|
|
||||||
- name: traefik
|
|
||||||
args:
|
|
||||||
- --entryPoints.metrics.address=:9100/tcp
|
|
||||||
- --entryPoints.traefik.address=:8080/tcp
|
|
||||||
- --entryPoints.web.address=:8000/tcp
|
|
||||||
- --entryPoints.websecure.address=:8443/tcp
|
|
||||||
- --entryPoints.pasarguard-api.address=:62051/tcp
|
|
||||||
- --api.dashboard=true
|
|
||||||
- --ping=true
|
|
||||||
- --metrics.prometheus=true
|
|
||||||
- --metrics.prometheus.entrypoint=metrics
|
|
||||||
- --providers.kubernetescrd
|
|
||||||
- --providers.kubernetescrd.allowEmptyServices=true
|
|
||||||
- --providers.kubernetesingress
|
|
||||||
- --providers.kubernetesingress.allowEmptyServices=true
|
|
||||||
- --providers.kubernetesingress.ingressendpoint.publishedservice=kube-system/traefik
|
|
||||||
- --entryPoints.websecure.http.tls=true
|
|
||||||
- --log.level=INFO
|
|
||||||
- --entryPoints.web.transport.respondingTimeouts.readTimeout=0s
|
|
||||||
- --entryPoints.websecure.transport.respondingTimeouts.readTimeout=0s
|
|
||||||
ports:
|
|
||||||
- containerPort: 9100
|
|
||||||
name: metrics
|
|
||||||
protocol: TCP
|
|
||||||
- containerPort: 8080
|
|
||||||
name: traefik
|
|
||||||
protocol: TCP
|
|
||||||
- containerPort: 8000
|
|
||||||
name: web
|
|
||||||
protocol: TCP
|
|
||||||
- containerPort: 8443
|
|
||||||
name: websecure
|
|
||||||
protocol: TCP
|
|
||||||
- containerPort: 62051
|
|
||||||
name: pasarguard-api
|
|
||||||
protocol: TCP
|
|
||||||
---
|
|
||||||
apiVersion: v1
|
|
||||||
kind: Service
|
|
||||||
metadata:
|
|
||||||
name: traefik
|
|
||||||
namespace: kube-system
|
|
||||||
spec:
|
|
||||||
ports:
|
|
||||||
- name: web
|
|
||||||
port: 80
|
|
||||||
protocol: TCP
|
|
||||||
targetPort: web
|
|
||||||
- name: websecure
|
|
||||||
port: 443
|
|
||||||
protocol: TCP
|
|
||||||
targetPort: websecure
|
|
||||||
- name: pasarguard-api
|
|
||||||
port: 62051
|
|
||||||
protocol: TCP
|
|
||||||
targetPort: pasarguard-api
|
|
||||||
@@ -16,14 +16,6 @@ helmCharts:
|
|||||||
valuesFile: syncthing-master.yaml
|
valuesFile: syncthing-master.yaml
|
||||||
includeCRDs: true
|
includeCRDs: true
|
||||||
|
|
||||||
- name: syncthing
|
|
||||||
repo: https://k8s-home-lab.github.io/helm-charts
|
|
||||||
version: 4.0.0
|
|
||||||
releaseName: syncthing-khv
|
|
||||||
namespace: syncthing
|
|
||||||
valuesFile: syncthing-khv.yaml
|
|
||||||
includeCRDs: true
|
|
||||||
|
|
||||||
- name: syncthing
|
- name: syncthing
|
||||||
repo: https://k8s-home-lab.github.io/helm-charts
|
repo: https://k8s-home-lab.github.io/helm-charts
|
||||||
version: 4.0.0
|
version: 4.0.0
|
||||||
@@ -31,3 +23,11 @@ helmCharts:
|
|||||||
namespace: syncthing
|
namespace: syncthing
|
||||||
valuesFile: syncthing-nas.yaml
|
valuesFile: syncthing-nas.yaml
|
||||||
includeCRDs: true
|
includeCRDs: true
|
||||||
|
|
||||||
|
# - name: syncthing
|
||||||
|
# repo: https://k8s-home-lab.github.io/helm-charts
|
||||||
|
# version: 4.0.0
|
||||||
|
# releaseName: syncthing-khv
|
||||||
|
# namespace: syncthing
|
||||||
|
# valuesFile: syncthing-khv.yaml
|
||||||
|
# includeCRDs: true
|
||||||
@@ -1,4 +1,3 @@
|
|||||||
---
|
|
||||||
apiVersion: apps/v1
|
apiVersion: apps/v1
|
||||||
kind: Deployment
|
kind: Deployment
|
||||||
metadata:
|
metadata:
|
||||||
@@ -23,7 +22,7 @@ spec:
|
|||||||
kubernetes.io/hostname: home.homenet
|
kubernetes.io/hostname: home.homenet
|
||||||
containers:
|
containers:
|
||||||
- name: desubot
|
- name: desubot
|
||||||
image: 'ultradesu/desubot:latest'
|
image: "ultradesu/desubot:latest"
|
||||||
imagePullPolicy: Always
|
imagePullPolicy: Always
|
||||||
envFrom:
|
envFrom:
|
||||||
- secretRef:
|
- secretRef:
|
||||||
@@ -36,7 +35,7 @@ spec:
|
|||||||
name: storage
|
name: storage
|
||||||
volumes:
|
volumes:
|
||||||
- name: storage
|
- name: storage
|
||||||
nfs:
|
persistentVolumeClaim:
|
||||||
server: nas.homenet
|
claimName: desubot-storage
|
||||||
path: /mnt/storage/Storage/k8s/desubot/
|
|
||||||
readOnly: false
|
readOnly: false
|
||||||
|
|
||||||
|
|||||||
@@ -8,3 +8,5 @@ resources:
|
|||||||
- external-secrets.yaml
|
- external-secrets.yaml
|
||||||
- desubot.yaml
|
- desubot.yaml
|
||||||
- restart-job.yaml
|
- restart-job.yaml
|
||||||
|
- storage.yaml
|
||||||
|
|
||||||
|
|||||||
12
k8s/apps/tg-bots/storage.yaml
Normal file
12
k8s/apps/tg-bots/storage.yaml
Normal file
@@ -0,0 +1,12 @@
|
|||||||
|
---
|
||||||
|
apiVersion: v1
|
||||||
|
kind: PersistentVolumeClaim
|
||||||
|
metadata:
|
||||||
|
name: desubot-storage
|
||||||
|
spec:
|
||||||
|
accessModes:
|
||||||
|
- ReadWriteMany
|
||||||
|
storageClassName: nfs-csi
|
||||||
|
resources:
|
||||||
|
requests:
|
||||||
|
storage: 200Gi
|
||||||
@@ -2,7 +2,7 @@
|
|||||||
|
|
||||||
global:
|
global:
|
||||||
domain: ag.hexor.cy
|
domain: ag.hexor.cy
|
||||||
nodeSelector:
|
nodeSelector: &nodeSelector
|
||||||
kubernetes.io/hostname: master.tail2fe2d.ts.net
|
kubernetes.io/hostname: master.tail2fe2d.ts.net
|
||||||
logging:
|
logging:
|
||||||
format: text
|
format: text
|
||||||
@@ -56,14 +56,14 @@ configs:
|
|||||||
controller:
|
controller:
|
||||||
replicas: 1
|
replicas: 1
|
||||||
nodeSelector:
|
nodeSelector:
|
||||||
kubernetes.io/hostname: master.tail2fe2d.ts.net
|
<<: *nodeSelector
|
||||||
# Add resources (requests/limits), PDB etc. if needed
|
# Add resources (requests/limits), PDB etc. if needed
|
||||||
|
|
||||||
# Dex OIDC provider
|
# Dex OIDC provider
|
||||||
dex:
|
dex:
|
||||||
replicas: 1
|
replicas: 1
|
||||||
nodeSelector:
|
nodeSelector:
|
||||||
kubernetes.io/hostname: master.tail2fe2d.ts.net
|
<<: *nodeSelector
|
||||||
enabled: false
|
enabled: false
|
||||||
|
|
||||||
# Standard Redis disabled because Redis HA is enabled
|
# Standard Redis disabled because Redis HA is enabled
|
||||||
@@ -86,7 +86,7 @@ redis-ha:
|
|||||||
server:
|
server:
|
||||||
replicas: 1
|
replicas: 1
|
||||||
nodeSelector:
|
nodeSelector:
|
||||||
kubernetes.io/hostname: master.tail2fe2d.ts.net
|
<<: *nodeSelector
|
||||||
ingress:
|
ingress:
|
||||||
enabled: false
|
enabled: false
|
||||||
|
|
||||||
@@ -99,8 +99,11 @@ server:
|
|||||||
# Repository Server
|
# Repository Server
|
||||||
repoServer:
|
repoServer:
|
||||||
replicas: 1
|
replicas: 1
|
||||||
|
livenessProbe:
|
||||||
|
timeoutSeconds: 10
|
||||||
|
periodSeconds: 60
|
||||||
nodeSelector:
|
nodeSelector:
|
||||||
kubernetes.io/hostname: master.tail2fe2d.ts.net
|
<<: *nodeSelector
|
||||||
# Add resources (requests/limits), PDB etc. if needed
|
# Add resources (requests/limits), PDB etc. if needed
|
||||||
|
|
||||||
# ApplicationSet Controller
|
# ApplicationSet Controller
|
||||||
@@ -108,7 +111,7 @@ applicationSet:
|
|||||||
enabled: true # Enabled by default
|
enabled: true # Enabled by default
|
||||||
replicas: 1
|
replicas: 1
|
||||||
nodeSelector:
|
nodeSelector:
|
||||||
kubernetes.io/hostname: master.tail2fe2d.ts.net
|
<<: *nodeSelector
|
||||||
# Add resources (requests/limits), PDB etc. if needed
|
# Add resources (requests/limits), PDB etc. if needed
|
||||||
|
|
||||||
# Notifications Controller
|
# Notifications Controller
|
||||||
@@ -116,5 +119,5 @@ notifications:
|
|||||||
enabled: true # Enabled by default
|
enabled: true # Enabled by default
|
||||||
replicas: 1
|
replicas: 1
|
||||||
nodeSelector:
|
nodeSelector:
|
||||||
kubernetes.io/hostname: master.tail2fe2d.ts.net
|
<<: *nodeSelector
|
||||||
# Add notifiers, triggers, templates configurations if needed
|
# Add notifiers, triggers, templates configurations if needed
|
||||||
|
|||||||
@@ -35,5 +35,6 @@ spec:
|
|||||||
key: secretKey
|
key: secretKey
|
||||||
selector:
|
selector:
|
||||||
dnsZones:
|
dnsZones:
|
||||||
|
- "ps.hexor.cy"
|
||||||
- "of.hexor.cy"
|
- "of.hexor.cy"
|
||||||
|
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
FROM debian:sid
|
FROM debian:sid
|
||||||
|
|
||||||
ENV BW_CLI_VERSION=2025.5.0
|
ENV BW_CLI_VERSION=2025.12.1
|
||||||
|
|
||||||
RUN apt update && \
|
RUN apt update && \
|
||||||
apt install -y wget unzip && \
|
apt install -y wget unzip && \
|
||||||
|
|||||||
@@ -37,15 +37,15 @@ spec:
|
|||||||
kubernetes.io/hostname: master.tail2fe2d.ts.net
|
kubernetes.io/hostname: master.tail2fe2d.ts.net
|
||||||
containers:
|
containers:
|
||||||
- name: bitwarden-cli
|
- name: bitwarden-cli
|
||||||
image: ultradesu/bitwarden-client:2025.5.0
|
image: ultradesu/bitwarden-client:2025.12.1
|
||||||
imagePullPolicy: Always
|
imagePullPolicy: Always
|
||||||
resources:
|
resources:
|
||||||
requests:
|
requests:
|
||||||
memory: "128Mi"
|
memory: "128Mi"
|
||||||
cpu: "100m"
|
cpu: "300m"
|
||||||
limits:
|
limits:
|
||||||
memory: "512Mi"
|
memory: "512Mi"
|
||||||
cpu: "500m"
|
cpu: "1000m"
|
||||||
env:
|
env:
|
||||||
- name: BW_HOST
|
- name: BW_HOST
|
||||||
valueFrom:
|
valueFrom:
|
||||||
|
|||||||
@@ -3,5 +3,15 @@ kind: Kustomization
|
|||||||
|
|
||||||
resources:
|
resources:
|
||||||
- app.yaml
|
- app.yaml
|
||||||
|
- nfs-storage.yaml
|
||||||
- coredns-internal-resolve.yaml
|
- coredns-internal-resolve.yaml
|
||||||
|
|
||||||
|
helmCharts:
|
||||||
|
- name: csi-driver-nfs
|
||||||
|
repo: https://raw.githubusercontent.com/kubernetes-csi/csi-driver-nfs/master/charts
|
||||||
|
version: 4.12.0
|
||||||
|
releaseName: csi-driver-nfs
|
||||||
|
namespace: kube-system
|
||||||
|
#valuesFile: values.yaml
|
||||||
|
includeCRDs: true
|
||||||
|
|
||||||
|
|||||||
14
k8s/core/kube-system-custom/nfs-storage.yaml
Normal file
14
k8s/core/kube-system-custom/nfs-storage.yaml
Normal file
@@ -0,0 +1,14 @@
|
|||||||
|
---
|
||||||
|
apiVersion: storage.k8s.io/v1
|
||||||
|
kind: StorageClass
|
||||||
|
metadata:
|
||||||
|
name: nfs-csi
|
||||||
|
provisioner: nfs.csi.k8s.io
|
||||||
|
parameters:
|
||||||
|
server: nas.homenet
|
||||||
|
share: /mnt/storage/Storage/PVC
|
||||||
|
reclaimPolicy: Retain
|
||||||
|
volumeBindingMode: Immediate
|
||||||
|
mountOptions:
|
||||||
|
- vers=4
|
||||||
|
- hard
|
||||||
@@ -125,6 +125,8 @@ spec:
|
|||||||
{{ .umami }}
|
{{ .umami }}
|
||||||
USER_mmdl: |-
|
USER_mmdl: |-
|
||||||
{{ .mmdl }}
|
{{ .mmdl }}
|
||||||
|
USER_n8n: |-
|
||||||
|
{{ .n8n }}
|
||||||
data:
|
data:
|
||||||
- secretKey: authentik
|
- secretKey: authentik
|
||||||
sourceRef:
|
sourceRef:
|
||||||
@@ -258,3 +260,15 @@ spec:
|
|||||||
metadataPolicy: None
|
metadataPolicy: None
|
||||||
key: 2a9deb39-ef22-433e-a1be-df1555625e22
|
key: 2a9deb39-ef22-433e-a1be-df1555625e22
|
||||||
property: fields[12].value
|
property: fields[12].value
|
||||||
|
- secretKey: n8n
|
||||||
|
sourceRef:
|
||||||
|
storeRef:
|
||||||
|
name: vaultwarden-login
|
||||||
|
kind: ClusterSecretStore
|
||||||
|
remoteRef:
|
||||||
|
conversionStrategy: Default
|
||||||
|
decodingStrategy: None
|
||||||
|
metadataPolicy: None
|
||||||
|
key: 2a9deb39-ef22-433e-a1be-df1555625e22
|
||||||
|
property: fields[13].value
|
||||||
|
|
||||||
|
|||||||
@@ -13,9 +13,6 @@ spec:
|
|||||||
targetRevision: HEAD
|
targetRevision: HEAD
|
||||||
path: k8s/core/prom-stack
|
path: k8s/core/prom-stack
|
||||||
syncPolicy:
|
syncPolicy:
|
||||||
automated:
|
|
||||||
selfHeal: true
|
|
||||||
prune: true
|
|
||||||
syncOptions:
|
syncOptions:
|
||||||
- CreateNamespace=true
|
- CreateNamespace=true
|
||||||
- ServerSideApply=true
|
- ServerSideApply=true
|
||||||
|
|||||||
@@ -79,3 +79,83 @@ spec:
|
|||||||
key: 2a9deb39-ef22-433e-a1be-df1555625e22
|
key: 2a9deb39-ef22-433e-a1be-df1555625e22
|
||||||
property: fields[2].value
|
property: fields[2].value
|
||||||
|
|
||||||
|
---
|
||||||
|
apiVersion: external-secrets.io/v1
|
||||||
|
kind: ExternalSecret
|
||||||
|
metadata:
|
||||||
|
name: alertmanager-telegram
|
||||||
|
spec:
|
||||||
|
target:
|
||||||
|
name: alertmanager-telegram-secret
|
||||||
|
deletionPolicy: Delete
|
||||||
|
template:
|
||||||
|
type: Opaque
|
||||||
|
data:
|
||||||
|
TELEGRAM_BOT_TOKEN: |-
|
||||||
|
{{ .bot_token }}
|
||||||
|
TELEGRAM_CHAT_ID: |-
|
||||||
|
{{ .chat_id }}
|
||||||
|
data:
|
||||||
|
- secretKey: bot_token
|
||||||
|
sourceRef:
|
||||||
|
storeRef:
|
||||||
|
name: vaultwarden-login
|
||||||
|
kind: ClusterSecretStore
|
||||||
|
remoteRef:
|
||||||
|
conversionStrategy: Default
|
||||||
|
decodingStrategy: None
|
||||||
|
metadataPolicy: None
|
||||||
|
key: eca0fb0b-3939-40a8-890a-6294863e5a65
|
||||||
|
property: fields[0].value
|
||||||
|
- secretKey: chat_id
|
||||||
|
sourceRef:
|
||||||
|
storeRef:
|
||||||
|
name: vaultwarden-login
|
||||||
|
kind: ClusterSecretStore
|
||||||
|
remoteRef:
|
||||||
|
conversionStrategy: Default
|
||||||
|
decodingStrategy: None
|
||||||
|
metadataPolicy: None
|
||||||
|
key: eca0fb0b-3939-40a8-890a-6294863e5a65
|
||||||
|
property: fields[1].value
|
||||||
|
|
||||||
|
---
|
||||||
|
apiVersion: external-secrets.io/v1
|
||||||
|
kind: ExternalSecret
|
||||||
|
metadata:
|
||||||
|
name: grafana-telegram
|
||||||
|
spec:
|
||||||
|
target:
|
||||||
|
name: grafana-telegram
|
||||||
|
deletionPolicy: Delete
|
||||||
|
template:
|
||||||
|
type: Opaque
|
||||||
|
data:
|
||||||
|
bot-token: |-
|
||||||
|
{{ .bot_token }}
|
||||||
|
chat-id: |-
|
||||||
|
{{ .chat_id }}
|
||||||
|
data:
|
||||||
|
- secretKey: bot_token
|
||||||
|
sourceRef:
|
||||||
|
storeRef:
|
||||||
|
name: vaultwarden-login
|
||||||
|
kind: ClusterSecretStore
|
||||||
|
remoteRef:
|
||||||
|
conversionStrategy: Default
|
||||||
|
decodingStrategy: None
|
||||||
|
metadataPolicy: None
|
||||||
|
key: eca0fb0b-3939-40a8-890a-6294863e5a65
|
||||||
|
property: fields[0].value
|
||||||
|
- secretKey: chat_id
|
||||||
|
sourceRef:
|
||||||
|
storeRef:
|
||||||
|
name: vaultwarden-login
|
||||||
|
kind: ClusterSecretStore
|
||||||
|
remoteRef:
|
||||||
|
conversionStrategy: Default
|
||||||
|
decodingStrategy: None
|
||||||
|
metadataPolicy: None
|
||||||
|
key: eca0fb0b-3939-40a8-890a-6294863e5a65
|
||||||
|
property: fields[1].value
|
||||||
|
|
||||||
|
|||||||
152
k8s/core/prom-stack/grafana-alerting-configmap.yaml
Normal file
152
k8s/core/prom-stack/grafana-alerting-configmap.yaml
Normal file
@@ -0,0 +1,152 @@
|
|||||||
|
apiVersion: v1
|
||||||
|
kind: ConfigMap
|
||||||
|
metadata:
|
||||||
|
name: grafana-alerting
|
||||||
|
namespace: prometheus
|
||||||
|
data:
|
||||||
|
rules.yaml: |
|
||||||
|
apiVersion: 1
|
||||||
|
groups:
|
||||||
|
- orgId: 1
|
||||||
|
name: pasarguard_alerts
|
||||||
|
folder: Kubernetes
|
||||||
|
interval: 1m
|
||||||
|
rules:
|
||||||
|
- uid: pasarguard_cpu_throttling
|
||||||
|
title: VPN CPU Throttle
|
||||||
|
condition: B
|
||||||
|
data:
|
||||||
|
- refId: A
|
||||||
|
relativeTimeRange:
|
||||||
|
from: 600
|
||||||
|
to: 0
|
||||||
|
datasourceUid: P76F38748CEC837F0
|
||||||
|
model:
|
||||||
|
expr: 'rate(container_cpu_cfs_throttled_periods_total{container="pasarguard-node"}[5m])'
|
||||||
|
refId: A
|
||||||
|
intervalMs: 1000
|
||||||
|
maxDataPoints: 43200
|
||||||
|
- refId: B
|
||||||
|
relativeTimeRange:
|
||||||
|
from: 600
|
||||||
|
to: 0
|
||||||
|
datasourceUid: __expr__
|
||||||
|
model:
|
||||||
|
conditions:
|
||||||
|
- evaluator:
|
||||||
|
params:
|
||||||
|
- 0.1
|
||||||
|
type: gt
|
||||||
|
operator:
|
||||||
|
type: and
|
||||||
|
query:
|
||||||
|
params: []
|
||||||
|
datasource:
|
||||||
|
type: __expr__
|
||||||
|
uid: __expr__
|
||||||
|
expression: A
|
||||||
|
reducer: last
|
||||||
|
refId: B
|
||||||
|
type: reduce
|
||||||
|
noDataState: NoData
|
||||||
|
execErrState: Alerting
|
||||||
|
for: 5m
|
||||||
|
annotations:
|
||||||
|
pod: '{{ $labels.pod }}'
|
||||||
|
node: '{{ $labels.node }}'
|
||||||
|
namespace: '{{ $labels.namespace }}'
|
||||||
|
throttle_rate: '{{ printf "%.2f" $values.A }}'
|
||||||
|
summary: 'VPN node throttling CPU'
|
||||||
|
labels:
|
||||||
|
severity: warning
|
||||||
|
|
||||||
|
- orgId: 1
|
||||||
|
name: kubernetes_alerts
|
||||||
|
folder: Kubernetes
|
||||||
|
interval: 30s
|
||||||
|
rules:
|
||||||
|
- uid: node_not_ready
|
||||||
|
title: Kubernetes Node Not Ready
|
||||||
|
condition: B
|
||||||
|
data:
|
||||||
|
- refId: A
|
||||||
|
relativeTimeRange:
|
||||||
|
from: 300
|
||||||
|
to: 0
|
||||||
|
datasourceUid: P76F38748CEC837F0
|
||||||
|
model:
|
||||||
|
expr: 'kube_node_status_condition{condition="Ready",status="true"} == 0'
|
||||||
|
refId: A
|
||||||
|
intervalMs: 1000
|
||||||
|
maxDataPoints: 43200
|
||||||
|
- refId: B
|
||||||
|
relativeTimeRange:
|
||||||
|
from: 300
|
||||||
|
to: 0
|
||||||
|
datasourceUid: __expr__
|
||||||
|
model:
|
||||||
|
conditions:
|
||||||
|
- evaluator:
|
||||||
|
params:
|
||||||
|
- 0
|
||||||
|
type: gt
|
||||||
|
operator:
|
||||||
|
type: and
|
||||||
|
query:
|
||||||
|
params: []
|
||||||
|
datasource:
|
||||||
|
type: __expr__
|
||||||
|
uid: __expr__
|
||||||
|
expression: A
|
||||||
|
reducer: last
|
||||||
|
refId: B
|
||||||
|
type: reduce
|
||||||
|
noDataState: Alerting
|
||||||
|
execErrState: Alerting
|
||||||
|
for: 0s
|
||||||
|
annotations:
|
||||||
|
node: '{{ $labels.node }}'
|
||||||
|
condition: '{{ $labels.condition }}'
|
||||||
|
summary: 'Kubernetes node is not ready'
|
||||||
|
labels:
|
||||||
|
severity: critical
|
||||||
|
|
||||||
|
contactpoints.yaml: |
|
||||||
|
apiVersion: 1
|
||||||
|
contactPoints:
|
||||||
|
- orgId: 1
|
||||||
|
name: telegram
|
||||||
|
receivers:
|
||||||
|
- uid: telegram_default
|
||||||
|
type: telegram
|
||||||
|
disableResolveMessage: false
|
||||||
|
settings:
|
||||||
|
bottoken: $TELEGRAM_BOT_TOKEN
|
||||||
|
chatid: "124317807"
|
||||||
|
message: |
|
||||||
|
{{ if eq .Status "firing" }}🔥 FIRING{{ else }}✅ RESOLVED{{ end }}
|
||||||
|
|
||||||
|
{{ range .Alerts }}
|
||||||
|
📊 <b>{{ .Labels.alertname }}</b>
|
||||||
|
{{ .Annotations.summary }}
|
||||||
|
|
||||||
|
{{ if .Annotations.node }}🖥 <b>Node:</b> <code>{{ .Annotations.node }}</code>{{ end }}
|
||||||
|
{{ if .Annotations.pod }}📦 <b>Pod:</b> <code>{{ .Annotations.pod }}</code>{{ end }}
|
||||||
|
{{ if .Annotations.namespace }}📁 <b>Namespace:</b> <code>{{ .Annotations.namespace }}</code>{{ end }}
|
||||||
|
{{ if .Annotations.throttle_rate }}⚠️ <b>Throttling rate:</b> {{ .Annotations.throttle_rate }}{{ end }}
|
||||||
|
|
||||||
|
🔗 <a href="{{ .GeneratorURL }}">View in Grafana</a>
|
||||||
|
{{ end }}
|
||||||
|
parse_mode: HTML
|
||||||
|
|
||||||
|
policies.yaml: |
|
||||||
|
apiVersion: 1
|
||||||
|
policies:
|
||||||
|
- orgId: 1
|
||||||
|
receiver: telegram
|
||||||
|
group_by:
|
||||||
|
- grafana_folder
|
||||||
|
- alertname
|
||||||
|
group_wait: 10s
|
||||||
|
group_interval: 5m
|
||||||
|
repeat_interval: 4h
|
||||||
@@ -38,6 +38,10 @@ datasources:
|
|||||||
url: http://prometheus-kube-prometheus-prometheus.prometheus.svc:9090
|
url: http://prometheus-kube-prometheus-prometheus.prometheus.svc:9090
|
||||||
access: proxy
|
access: proxy
|
||||||
isDefault: true
|
isDefault: true
|
||||||
|
- name: Loki
|
||||||
|
type: loki
|
||||||
|
url: http://loki-gateway.prometheus.svc:80
|
||||||
|
access: proxy
|
||||||
|
|
||||||
ingress:
|
ingress:
|
||||||
enabled: true
|
enabled: true
|
||||||
@@ -52,3 +56,30 @@ ingress:
|
|||||||
hosts:
|
hosts:
|
||||||
- '*.hexor.cy'
|
- '*.hexor.cy'
|
||||||
|
|
||||||
|
extraConfigmapMounts:
|
||||||
|
- name: grafana-alerting-rules
|
||||||
|
mountPath: /etc/grafana/provisioning/alerting/rules.yaml
|
||||||
|
configMap: grafana-alerting
|
||||||
|
subPath: rules.yaml
|
||||||
|
readOnly: true
|
||||||
|
- name: grafana-alerting-contactpoints
|
||||||
|
mountPath: /etc/grafana/provisioning/alerting/contactpoints.yaml
|
||||||
|
configMap: grafana-alerting
|
||||||
|
subPath: contactpoints.yaml
|
||||||
|
readOnly: true
|
||||||
|
- name: grafana-alerting-policies
|
||||||
|
mountPath: /etc/grafana/provisioning/alerting/policies.yaml
|
||||||
|
configMap: grafana-alerting
|
||||||
|
subPath: policies.yaml
|
||||||
|
readOnly: true
|
||||||
|
|
||||||
|
envValueFrom:
|
||||||
|
TELEGRAM_BOT_TOKEN:
|
||||||
|
secretKeyRef:
|
||||||
|
name: grafana-telegram
|
||||||
|
key: bot-token
|
||||||
|
TELEGRAM_CHAT_ID:
|
||||||
|
secretKeyRef:
|
||||||
|
name: grafana-telegram
|
||||||
|
key: chat-id
|
||||||
|
|
||||||
|
|||||||
@@ -2,9 +2,9 @@ apiVersion: kustomize.config.k8s.io/v1beta1
|
|||||||
kind: Kustomization
|
kind: Kustomization
|
||||||
|
|
||||||
resources:
|
resources:
|
||||||
- app.yaml
|
|
||||||
- persistentVolume.yaml
|
- persistentVolume.yaml
|
||||||
- external-secrets.yaml
|
- external-secrets.yaml
|
||||||
|
- grafana-alerting-configmap.yaml
|
||||||
|
|
||||||
helmCharts:
|
helmCharts:
|
||||||
- name: kube-prometheus-stack
|
- name: kube-prometheus-stack
|
||||||
@@ -23,3 +23,18 @@ helmCharts:
|
|||||||
valuesFile: grafana-values.yaml
|
valuesFile: grafana-values.yaml
|
||||||
includeCRDs: true
|
includeCRDs: true
|
||||||
|
|
||||||
|
- name: loki
|
||||||
|
repo: https://grafana.github.io/helm-charts
|
||||||
|
version: 6.29.0
|
||||||
|
releaseName: loki
|
||||||
|
namespace: prometheus
|
||||||
|
valuesFile: loki-values.yaml
|
||||||
|
includeCRDs: true
|
||||||
|
|
||||||
|
- name: promtail
|
||||||
|
repo: https://grafana.github.io/helm-charts
|
||||||
|
version: 6.16.6
|
||||||
|
releaseName: promtail
|
||||||
|
namespace: prometheus
|
||||||
|
valuesFile: promtail-values.yaml
|
||||||
|
|
||||||
|
|||||||
75
k8s/core/prom-stack/loki-values.yaml
Normal file
75
k8s/core/prom-stack/loki-values.yaml
Normal file
@@ -0,0 +1,75 @@
|
|||||||
|
# Loki SingleBinary mode - optimal for homelab
|
||||||
|
deploymentMode: SingleBinary
|
||||||
|
|
||||||
|
loki:
|
||||||
|
auth_enabled: false
|
||||||
|
commonConfig:
|
||||||
|
replication_factor: 1
|
||||||
|
path_prefix: /var/loki
|
||||||
|
schemaConfig:
|
||||||
|
configs:
|
||||||
|
- from: 2024-01-01
|
||||||
|
store: tsdb
|
||||||
|
object_store: filesystem
|
||||||
|
schema: v13
|
||||||
|
index:
|
||||||
|
prefix: index_
|
||||||
|
period: 24h
|
||||||
|
storage:
|
||||||
|
type: filesystem
|
||||||
|
filesystem:
|
||||||
|
chunks_directory: /var/loki/chunks
|
||||||
|
rules_directory: /var/loki/rules
|
||||||
|
limits_config:
|
||||||
|
reject_old_samples: false
|
||||||
|
ingestion_rate_mb: 16
|
||||||
|
ingestion_burst_size_mb: 32
|
||||||
|
max_query_parallelism: 32
|
||||||
|
volume_enabled: true
|
||||||
|
|
||||||
|
singleBinary:
|
||||||
|
replicas: 1
|
||||||
|
nodeSelector:
|
||||||
|
kubernetes.io/hostname: master.tail2fe2d.ts.net
|
||||||
|
persistence:
|
||||||
|
enabled: true
|
||||||
|
size: 50Gi
|
||||||
|
storageClass: ""
|
||||||
|
|
||||||
|
# Disable distributed mode components
|
||||||
|
read:
|
||||||
|
replicas: 0
|
||||||
|
write:
|
||||||
|
replicas: 0
|
||||||
|
backend:
|
||||||
|
replicas: 0
|
||||||
|
|
||||||
|
# Disable memcached (not needed for SingleBinary)
|
||||||
|
chunksCache:
|
||||||
|
enabled: false
|
||||||
|
resultsCache:
|
||||||
|
enabled: false
|
||||||
|
|
||||||
|
# Gateway for Loki access
|
||||||
|
gateway:
|
||||||
|
enabled: true
|
||||||
|
replicas: 1
|
||||||
|
service:
|
||||||
|
type: ClusterIP
|
||||||
|
|
||||||
|
# Disable tests and canary
|
||||||
|
test:
|
||||||
|
enabled: false
|
||||||
|
lokiCanary:
|
||||||
|
enabled: false
|
||||||
|
|
||||||
|
# Monitoring
|
||||||
|
monitoring:
|
||||||
|
dashboards:
|
||||||
|
enabled: false
|
||||||
|
rules:
|
||||||
|
enabled: false
|
||||||
|
serviceMonitor:
|
||||||
|
enabled: false
|
||||||
|
selfMonitoring:
|
||||||
|
enabled: false
|
||||||
@@ -1,5 +1,35 @@
|
|||||||
grafana:
|
grafana:
|
||||||
enabled: false
|
enabled: false
|
||||||
|
|
||||||
|
alertmanager:
|
||||||
|
config:
|
||||||
|
global:
|
||||||
|
telegram_api_url: "https://api.telegram.org"
|
||||||
|
route:
|
||||||
|
group_by: ['alertname', 'cluster', 'service']
|
||||||
|
group_wait: 10s
|
||||||
|
group_interval: 10s
|
||||||
|
repeat_interval: 12h
|
||||||
|
receiver: 'telegram'
|
||||||
|
receivers:
|
||||||
|
- name: 'telegram'
|
||||||
|
telegram_configs:
|
||||||
|
- bot_token: '${TELEGRAM_BOT_TOKEN}'
|
||||||
|
chat_id: ${TELEGRAM_CHAT_ID}
|
||||||
|
parse_mode: 'HTML'
|
||||||
|
message: |
|
||||||
|
{{ range .Alerts }}
|
||||||
|
<b>{{ .Labels.alertname }}</b>
|
||||||
|
{{ if .Labels.severity }}<b>Severity:</b> {{ .Labels.severity }}{{ end }}
|
||||||
|
<b>Status:</b> {{ .Status }}
|
||||||
|
{{ if .Annotations.summary }}<b>Summary:</b> {{ .Annotations.summary }}{{ end }}
|
||||||
|
{{ if .Annotations.description }}<b>Description:</b> {{ .Annotations.description }}{{ end }}
|
||||||
|
{{ end }}
|
||||||
|
|
||||||
|
alertmanagerSpec:
|
||||||
|
secrets:
|
||||||
|
- alertmanager-telegram-secret
|
||||||
|
|
||||||
prometheus:
|
prometheus:
|
||||||
prometheusSpec:
|
prometheusSpec:
|
||||||
enableRemoteWriteReceiver: true
|
enableRemoteWriteReceiver: true
|
||||||
|
|||||||
37
k8s/core/prom-stack/promtail-values.yaml
Normal file
37
k8s/core/prom-stack/promtail-values.yaml
Normal file
@@ -0,0 +1,37 @@
|
|||||||
|
# Promtail - log collection agent for all cluster pods
|
||||||
|
config:
|
||||||
|
clients:
|
||||||
|
- url: http://loki-gateway.prometheus.svc:80/loki/api/v1/push
|
||||||
|
|
||||||
|
# DaemonSet - runs on every node
|
||||||
|
daemonset:
|
||||||
|
enabled: true
|
||||||
|
|
||||||
|
# Tolerations for master/control-plane nodes
|
||||||
|
tolerations:
|
||||||
|
- key: node-role.kubernetes.io/master
|
||||||
|
operator: Exists
|
||||||
|
effect: NoSchedule
|
||||||
|
- key: node-role.kubernetes.io/control-plane
|
||||||
|
operator: Exists
|
||||||
|
effect: NoSchedule
|
||||||
|
|
||||||
|
# Init container to increase inotify limits
|
||||||
|
initContainer:
|
||||||
|
- name: init-inotify
|
||||||
|
image: docker.io/busybox:1.36
|
||||||
|
imagePullPolicy: IfNotPresent
|
||||||
|
command:
|
||||||
|
- sh
|
||||||
|
- -c
|
||||||
|
- sysctl -w fs.inotify.max_user_instances=512
|
||||||
|
securityContext:
|
||||||
|
privileged: true
|
||||||
|
|
||||||
|
resources:
|
||||||
|
requests:
|
||||||
|
cpu: 50m
|
||||||
|
memory: 64Mi
|
||||||
|
limits:
|
||||||
|
cpu: 200m
|
||||||
|
memory: 128Mi
|
||||||
@@ -16,7 +16,7 @@ spec:
|
|||||||
serviceAccountName: system-upgrade
|
serviceAccountName: system-upgrade
|
||||||
upgrade:
|
upgrade:
|
||||||
image: rancher/k3s-upgrade
|
image: rancher/k3s-upgrade
|
||||||
version: v1.34.2+k3s1
|
version: v1.34.3+k3s1
|
||||||
---
|
---
|
||||||
# Agent plan
|
# Agent plan
|
||||||
apiVersion: upgrade.cattle.io/v1
|
apiVersion: upgrade.cattle.io/v1
|
||||||
@@ -39,5 +39,5 @@ spec:
|
|||||||
serviceAccountName: system-upgrade
|
serviceAccountName: system-upgrade
|
||||||
upgrade:
|
upgrade:
|
||||||
image: rancher/k3s-upgrade
|
image: rancher/k3s-upgrade
|
||||||
version: v1.34.2+k3s1
|
version: v1.34.3+k3s1
|
||||||
|
|
||||||
|
|||||||
@@ -47,6 +47,11 @@ spec:
|
|||||||
spec:
|
spec:
|
||||||
nodeSelector:
|
nodeSelector:
|
||||||
kubernetes.io/hostname: uk-desktop.tail2fe2d.ts.net
|
kubernetes.io/hostname: uk-desktop.tail2fe2d.ts.net
|
||||||
|
tolerations:
|
||||||
|
- key: workload
|
||||||
|
operator: Equal
|
||||||
|
value: desktop
|
||||||
|
effect: NoSchedule
|
||||||
volumes:
|
volumes:
|
||||||
- name: config
|
- name: config
|
||||||
hostPath:
|
hostPath:
|
||||||
@@ -97,3 +102,22 @@ spec:
|
|||||||
port: 80
|
port: 80
|
||||||
targetPort: 8080
|
targetPort: 8080
|
||||||
|
|
||||||
|
---
|
||||||
|
apiVersion: networking.k8s.io/v1
|
||||||
|
kind: Ingress
|
||||||
|
metadata:
|
||||||
|
name: jf-local-ingress
|
||||||
|
annotations:
|
||||||
|
ingressClassName: traefik
|
||||||
|
spec:
|
||||||
|
rules:
|
||||||
|
- host: tr.uk
|
||||||
|
http:
|
||||||
|
paths:
|
||||||
|
- path: /
|
||||||
|
pathType: Prefix
|
||||||
|
backend:
|
||||||
|
service:
|
||||||
|
name: qbittorrent
|
||||||
|
port:
|
||||||
|
number: 80
|
||||||
|
|||||||
@@ -9,6 +9,11 @@ resources:
|
|||||||
cpu: "6000m"
|
cpu: "6000m"
|
||||||
nodeSelector:
|
nodeSelector:
|
||||||
kubernetes.io/hostname: uk-desktop.tail2fe2d.ts.net
|
kubernetes.io/hostname: uk-desktop.tail2fe2d.ts.net
|
||||||
|
tolerations:
|
||||||
|
- key: workload
|
||||||
|
operator: Equal
|
||||||
|
value: desktop
|
||||||
|
effect: NoSchedule
|
||||||
persistence:
|
persistence:
|
||||||
config:
|
config:
|
||||||
enabled: true
|
enabled: true
|
||||||
@@ -28,11 +33,8 @@ persistence:
|
|||||||
ingress:
|
ingress:
|
||||||
enabled: true
|
enabled: true
|
||||||
className: traefik
|
className: traefik
|
||||||
annotations:
|
|
||||||
cert-manager.io/cluster-issuer: letsencrypt
|
|
||||||
traefik.ingress.kubernetes.io/router.middlewares: kube-system-https-redirect@kubernetescrd
|
|
||||||
hosts:
|
hosts:
|
||||||
- host: uk-desktop.uk
|
- host: jf.uk
|
||||||
paths:
|
paths:
|
||||||
- path: /
|
- path: /
|
||||||
pathType: Prefix
|
pathType: Prefix
|
||||||
|
|||||||
@@ -1,4 +1,3 @@
|
|||||||
|
|
||||||
data "authentik_flow" "default_authorization_flow" {
|
data "authentik_flow" "default_authorization_flow" {
|
||||||
slug = var.default_authorization_flow
|
slug = var.default_authorization_flow
|
||||||
}
|
}
|
||||||
@@ -299,7 +298,7 @@ resource "authentik_outpost" "outposts" {
|
|||||||
kubernetes_ingress_class_name = null
|
kubernetes_ingress_class_name = null
|
||||||
kubernetes_disabled_components = []
|
kubernetes_disabled_components = []
|
||||||
kubernetes_ingress_annotations = {}
|
kubernetes_ingress_annotations = {}
|
||||||
kubernetes_ingress_secret_name = "authentik-outpost-tls"
|
kubernetes_ingress_secret_name = "idm-tls"
|
||||||
})
|
})
|
||||||
|
|
||||||
depends_on = [
|
depends_on = [
|
||||||
|
|||||||
@@ -51,6 +51,9 @@ proxy_applications = {
|
|||||||
internal_host = "http://secret-reader.k8s-secret.svc:80"
|
internal_host = "http://secret-reader.k8s-secret.svc:80"
|
||||||
internal_host_ssl_validation = false
|
internal_host_ssl_validation = false
|
||||||
meta_description = ""
|
meta_description = ""
|
||||||
|
skip_path_regex = <<-EOT
|
||||||
|
/webhook
|
||||||
|
EOT
|
||||||
meta_icon = "https://img.icons8.com/ios-filled/50/password.png"
|
meta_icon = "https://img.icons8.com/ios-filled/50/password.png"
|
||||||
mode = "proxy"
|
mode = "proxy"
|
||||||
outpost = "kubernetes-outpost"
|
outpost = "kubernetes-outpost"
|
||||||
|
|||||||
Reference in New Issue
Block a user