Compare commits
1 Commits
main
...
caddy-dock
Author | SHA1 | Date |
---|---|---|
![]() |
d107193958 | 2 years ago |
@ -1,3 +1,4 @@
|
||||
/secrets/
|
||||
/data/
|
||||
env.production
|
||||
secrets/
|
||||
data/
|
||||
.redo
|
||||
*.tmp
|
||||
|
@ -0,0 +1 @@
|
||||
DOMAIN=localhost
|
@ -1,15 +0,0 @@
|
||||
# copy this to .env and it will be sourced by the appropriate services
|
||||
|
||||
# domain your services will be running on
|
||||
DOMAIN=localhost
|
||||
|
||||
# admin user for auth
|
||||
ADMIN_USER=
|
||||
ADMIN_PASS=
|
||||
|
||||
# used for sending notifications and reset passwords
|
||||
# only supports smtp+starttls
|
||||
SMTP_ADDR=
|
||||
SMTP_PORT=587
|
||||
SMTP_USER=
|
||||
SMTP_PASS=
|
@ -1,8 +0,0 @@
|
||||
podman compose --env-file ${ENV_FILE:-.env} \
|
||||
--file services/secrets.yaml \
|
||||
--file services/backup.yaml \
|
||||
--file services/proxy.yaml \
|
||||
--file services/auth.yaml \
|
||||
--file services/web.yaml \
|
||||
--file services/git.yaml \
|
||||
down --volumes
|
@ -1,8 +0,0 @@
|
||||
podman compose --env-file ${ENV_FILE:-.env} \
|
||||
--file services/secrets.yaml \
|
||||
--file services/backup.yaml \
|
||||
--file services/proxy.yaml \
|
||||
--file services/auth.yaml \
|
||||
--file services/web.yaml \
|
||||
--file services/git.yaml \
|
||||
exec "$@"
|
@ -1,4 +0,0 @@
|
||||
echo generating zitadel secrets; {
|
||||
openssl rand -hex 16 | tr -d '\n' >! secrets/auth/zitadel/MASTER_KEY
|
||||
openssl rand -hex 32 | tr -d '\n' >! secrets/auth/zitadel/STORAGE_PASSWORD
|
||||
}
|
@ -1,8 +0,0 @@
|
||||
podman compose --env-file ${ENV_FILE:-.env} \
|
||||
--file services/secrets.yaml \
|
||||
--file services/backup.yaml \
|
||||
--file services/proxy.yaml \
|
||||
--file services/auth.yaml \
|
||||
--file services/web.yaml \
|
||||
--file services/git.yaml \
|
||||
ps
|
@ -1,8 +0,0 @@
|
||||
podman compose --env-file ${ENV_FILE:-.env} \
|
||||
--file services/secrets.yaml \
|
||||
--file services/backup.yaml \
|
||||
--file services/proxy.yaml \
|
||||
--file services/auth.yaml \
|
||||
--file services/web.yaml \
|
||||
--file services/git.yaml \
|
||||
pull
|
@ -1,8 +0,0 @@
|
||||
podman compose --env-file ${ENV_FILE:-.env} \
|
||||
--file services/secrets.yaml \
|
||||
--file services/backup.yaml \
|
||||
--file services/proxy.yaml \
|
||||
--file services/auth.yaml \
|
||||
--file services/web.yaml \
|
||||
--file services/git.yaml \
|
||||
run "$@"
|
@ -1,8 +0,0 @@
|
||||
podman compose --env-file ${ENV_FILE:-.env} \
|
||||
--file services/secrets.yaml \
|
||||
--file services/backup.yaml \
|
||||
--file services/proxy.yaml \
|
||||
--file services/auth.yaml \
|
||||
--file services/web.yaml \
|
||||
--file services/git.yaml \
|
||||
up --build
|
@ -1,69 +0,0 @@
|
||||
secrets:
|
||||
MASTER_KEY:
|
||||
file: ../secrets/auth/zitadel/MASTER_KEY
|
||||
|
||||
services:
|
||||
backup:
|
||||
volumes:
|
||||
- ../data/auth:/mnt/backup/src/auth:ro
|
||||
|
||||
generate-secrets:
|
||||
volumes:
|
||||
- ../secrets/auth/zitadel/MASTER_KEY:/secrets/auth/zitadel/MASTER_KEY
|
||||
|
||||
zitadel:
|
||||
restart: 'unless-stopped'
|
||||
image: 'ghcr.io/zitadel/zitadel:v2.48.3'
|
||||
environment:
|
||||
ZITADEL_DATABASE_COCKROACH_HOST: crdb
|
||||
ZITADEL_EXTERNALSECURE: true
|
||||
ZITADEL_EXTERNALDOMAIN: auth.${DOMAIN}
|
||||
ZITADEL_EXTERNALPORT: 443
|
||||
ZITADEL_WEBAUTHN_NAME: ${DOMAIN}
|
||||
ZITADEL_FIRSTINSTANCE_ORG_NAME: basement
|
||||
ZITADEL_FIRSTINSTANCE_ORG_HUMAN_USERNAME: ${ADMIN_USER}
|
||||
ZITADEL_FIRSTINSTANCE_ORG_HUMAN_PASSWORD: ${ADMIN_PASS}
|
||||
|
||||
ZITADEL_DEFAULTINSTANCE_SMTPCONFIGURATION_SMTP_HOST: "${SMTP_ADDR}:${SMTP_PORT}"
|
||||
ZITADEL_DEFAULTINSTANCE_SMTPCONFIGURATION_SMTP_USER: ${SMTP_USER}
|
||||
ZITADEL_DEFAULTINSTANCE_SMTPCONFIGURATION_SMTP_PASSWORD: ${SMTP_PASS}
|
||||
ZITADEL_DEFAULTINSTANCE_SMTPCONFIGURATION_SMTP_SSL: true
|
||||
|
||||
ZITADEL_DEFAULTINSTANCE_SMTPCONFIGURATION_FROM: basement@mail.${DOMAIN}
|
||||
ZITADEL_DEFAULTINSTANCE_SMTPCONFIGURATION_FROMNAME: basement
|
||||
ZITADEL_DEFAULTINSTANCE_SMTPCONFIGURATION_SMTP_REPLYTOADDRESS: basement@mail.${DOMAIN}
|
||||
secrets:
|
||||
- MASTER_KEY
|
||||
command: "start-from-init --masterkeyFile /run/secrets/MASTER_KEY --tlsMode external"
|
||||
depends_on:
|
||||
generate-secrets:
|
||||
condition: 'service_completed_successfully'
|
||||
caddy:
|
||||
condition: 'service_healthy'
|
||||
crdb:
|
||||
condition: 'service_healthy'
|
||||
ports:
|
||||
- '8080:8080'
|
||||
|
||||
crdb:
|
||||
restart: unless-stopped
|
||||
image: 'cockroachdb/cockroach:latest-v23.1'
|
||||
depends_on:
|
||||
generate-secrets:
|
||||
condition: 'service_completed_successfully'
|
||||
command: "start-single-node --insecure --store=path=/cockroach/cockroach-data,size=20%"
|
||||
healthcheck:
|
||||
test: ["CMD", "curl", "--fail", "http://localhost:8080/health?ready=1"]
|
||||
interval: '10s'
|
||||
timeout: '30s'
|
||||
retries: 5
|
||||
start_period: '20s'
|
||||
ports:
|
||||
- '9090:8080'
|
||||
- '26257:26257'
|
||||
volumes:
|
||||
- ../data/auth/crdb/data:/cockroach/cockroach-data:rw
|
||||
|
||||
caddy:
|
||||
volumes:
|
||||
- ./auth/Proxyfile:/etc/caddy.d/zitadel:ro
|
@ -1,4 +0,0 @@
|
||||
auth.{$DOMAIN}:443 {
|
||||
reverse_proxy zitadel:8080
|
||||
tls internal
|
||||
}
|
@ -1,42 +0,0 @@
|
||||
secrets:
|
||||
B2_APPLICATION_KEY:
|
||||
file: ../secrets/backup/duplicity/B2_APPLICATION_KEY
|
||||
B2_APPLICATION_KEY_ID:
|
||||
file: ../secrets/backup/duplicity/B2_APPLICATION_KEY_ID
|
||||
BUCKET_NAME:
|
||||
file: ../secrets/backup/duplicity/BUCKET_NAME
|
||||
PASSPHRASE:
|
||||
file: ../secrets/backup/duplicity/PASSPHRASE
|
||||
|
||||
services:
|
||||
backup:
|
||||
image: ghcr.io/tecnativa/docker-duplicity:3.3.1
|
||||
restart: unless-stopped
|
||||
depends_on:
|
||||
generate-secrets:
|
||||
condition: 'service_completed_successfully'
|
||||
secrets: [B2_APPLICATION_KEY, B2_APPLICATION_KEY_ID, BUCKET_NAME, PASSPHRASE]
|
||||
environment:
|
||||
HOSTNAME: ${DOMAIN}
|
||||
TZ: America/New_York
|
||||
volumes:
|
||||
- ./backup/backup-files:/backup-files:ro
|
||||
entrypoint: ["/bin/sh", "/backup-files"]
|
||||
|
||||
generate-secrets:
|
||||
volumes:
|
||||
- ../secrets/backup/duplicity/BUCKET_NAME:/secrets/backup/duplicity/BUCKET_NAME
|
||||
- ../secrets/backup/duplicity/PASSPHRASE:/secrets/backup/duplicity/PASSPHRASE
|
||||
|
||||
|
||||
# duplicity-postgres:
|
||||
# image: tecnativa/docker-duplicity-postgres:latest
|
||||
# restart: unless-stopped
|
||||
# depends_on: [secrets]
|
||||
# secrets: [B2_APPLICATION_KEY, B2_APPLICATION_KEY_ID, BUCKET_NAME, PASSPHRASE]
|
||||
# environment:
|
||||
# HOSTNAME: ${DOMAIN}
|
||||
# TZ: America/New_York
|
||||
# volumes:
|
||||
# - ./backup/backup-databases:/backup-databases:ro
|
||||
# entrypoint: ["/bin/sh", "/backup-databases"]
|
@ -1,14 +0,0 @@
|
||||
read B2_APPLICATION_KEY_ID < /run/secrets/B2_APPLICATION_KEY_ID
|
||||
read B2_APPLICATION_KEY < /run/secrets/B2_APPLICATION_KEY
|
||||
read BUCKET_NAME < /run/secrets/BUCKET_NAME
|
||||
export DST=b2://${B2_APPLICATION_KEY_ID}:${B2_APPLICATION_KEY}@${BUCKET_NAME}
|
||||
|
||||
read PASSPHRASE < /run/secrets/PASSPHRASE
|
||||
export PASSPHRASE
|
||||
|
||||
for environment in /backup/*; do
|
||||
. $environment
|
||||
export PGHOST PGPASSWORD PGUSER DBS_TO_INCLUDE DBS_TO_EXCLUDE
|
||||
/usr/local/bin/entrypoint
|
||||
unset PGHOST PGPASSWORD PGUSER DBS_TO_INCLUDE DBS_TO_EXCLUDE
|
||||
done
|
@ -1,9 +0,0 @@
|
||||
read B2_APPLICATION_KEY_ID < /run/secrets/B2_APPLICATION_KEY_ID
|
||||
read B2_APPLICATION_KEY < /run/secrets/B2_APPLICATION_KEY
|
||||
read BUCKET_NAME < /run/secrets/BUCKET_NAME
|
||||
export DST=b2://${B2_APPLICATION_KEY_ID}:${B2_APPLICATION_KEY}@${BUCKET_NAME}
|
||||
|
||||
read PASSPHRASE < /run/secrets/PASSPHRASE
|
||||
export PASSPHRASE
|
||||
|
||||
/usr/local/bin/entrypoint
|
@ -0,0 +1,14 @@
|
||||
. ../../env.production
|
||||
|
||||
service=$(basename $PWD)
|
||||
secrets="../../secrets/$service"
|
||||
|
||||
read B2_APPLICATION_KEY_ID < $secrets/application-key-id
|
||||
read B2_APPLICATION_KEY < $secrets/application-key
|
||||
export BUCKET_NAME=${DOMAIN}-backup
|
||||
|
||||
export DESTINATION=b2://${B2_APPLICATION_KEY_ID}:${B2_APPLICATION_KEY}@${BUCKET_NAME}
|
||||
|
||||
read PASSPHRASE < $secrets/passphrase
|
||||
env PASSPHRASE=$PASSPHRASE duplicity backup ../../data $DESTINATION >&2
|
||||
env PASSPHRASE=$PASSPHRASE duplicity remove-older-than 28D $DESTINATION >&2
|
@ -0,0 +1,32 @@
|
||||
version: "3.7"
|
||||
|
||||
services:
|
||||
caddy:
|
||||
image: lucaslorentz/caddy-docker-proxy:ci-alpine
|
||||
restart: unless-stopped
|
||||
ports:
|
||||
- "80:80"
|
||||
- "443:443"
|
||||
- "443:443/udp"
|
||||
privileged: true
|
||||
volumes:
|
||||
- /var/run/docker.sock:/var/run/docker.sock:ro
|
||||
- ./caddy/Caddyfile:/etc/caddy/Caddyfile
|
||||
- ../data/caddy/site:/site
|
||||
- ../data/caddy/data:/data
|
||||
- caddy_config:/config
|
||||
environment:
|
||||
- DOMAIN
|
||||
- CADDY_INGRESS_NETWORKS=caddy
|
||||
labels:
|
||||
caddy: ${DOMAIN}
|
||||
caddy.file_server.root: /site
|
||||
networks:
|
||||
- caddy
|
||||
|
||||
networks:
|
||||
caddy:
|
||||
external: true
|
||||
|
||||
volumes:
|
||||
caddy_config:
|
@ -1,48 +0,0 @@
|
||||
secrets:
|
||||
DB_PASSWD:
|
||||
file: ../secrets/git/gitea/DB_PASSWD
|
||||
|
||||
services:
|
||||
caddy:
|
||||
volumes:
|
||||
- ./git/Proxyfile:/etc/caddy.d/git
|
||||
backup:
|
||||
volumes:
|
||||
- ../data/git:/mnt/backup/src/git
|
||||
|
||||
gitea:
|
||||
image: gitea/gitea:1.21.3-rootless
|
||||
secrets: [ DB_PASSWD ]
|
||||
environment:
|
||||
GITEA__database__DB_TYPE: postgres
|
||||
GITEA__database__HOST: "db:5432"
|
||||
GITEA__database__NAME: gitea
|
||||
GITEA__database__USER: gitea
|
||||
GITEA__database__PASSWD__FILE: /run/secrets/DB_PASSWD
|
||||
GITEA__mailer__ENABLED: true
|
||||
GITEA__mailer__FROM: gitea@mail.${DOMAIN}
|
||||
GITEA__mailer__PROTOCOL: smtp+starttls
|
||||
GITEA__mailer__SMTP_ADDR: ${SMTP_ADDR}
|
||||
GITEA__mailer__SMTP_PORT: ${SMTP_PORT}
|
||||
GITEA__mailer__USER: ${SMTP_USER}
|
||||
GITEA__mailer__PASSWD: ${SMTP_PASS}
|
||||
restart: unless-stopped
|
||||
volumes:
|
||||
- ../data/git/gitea/data:/data
|
||||
ports:
|
||||
- 3000:3000
|
||||
db:
|
||||
image: postgres:16.1-alpine
|
||||
secrets: [ DB_PASSWD ]
|
||||
environment:
|
||||
POSTGRES_USER: gitea
|
||||
POSTGRES_PASSWORD_FILE: /run/secrets/DB_PASSWD
|
||||
POSTGRES_DB: gitea
|
||||
restart: unless-stopped
|
||||
volumes:
|
||||
- db_data:/var/lib/postgresql/data
|
||||
expose:
|
||||
- 5432
|
||||
|
||||
volumes:
|
||||
db_data:
|
@ -1,3 +0,0 @@
|
||||
git.{$DOMAIN} {
|
||||
reverse_proxy gitea:3000
|
||||
}
|
@ -1,54 +0,0 @@
|
||||
secrets:
|
||||
SMTP_PASSWORD:
|
||||
file: ../secrets/mail/SMTP_PASSWORD
|
||||
|
||||
services:
|
||||
generate-secrets:
|
||||
volumes:
|
||||
- ../secrets/mail/maddy/SMTP_PASSWORD:/secrets/mail/maddy/SMTP_PASSWORD
|
||||
|
||||
backup:
|
||||
volumes:
|
||||
- ../data/mail:/mnt/backup/src/mail:ro
|
||||
|
||||
caddy:
|
||||
volumes:
|
||||
- ./mail/Proxyfile:/etc/caddy.d/mail:ro
|
||||
|
||||
maddy:
|
||||
image: foxcpp/maddy:0.7
|
||||
secrets: [SMTP_PASSWORD]
|
||||
restart: unless-stopped
|
||||
depends_on:
|
||||
generate-secrets:
|
||||
condition: 'service_completed_successfully'
|
||||
environment:
|
||||
- MADDY_HOSTNAME=mx.mail.${DOMAIN}
|
||||
- MADDY_DOMAIN=mail.${DOMAIN}
|
||||
volumes:
|
||||
- ../data/mail/maddy:/data
|
||||
# TODO: get from caddy?
|
||||
#- ../secrets/tls/fullchain.pem:/data/tls/fullchain.pem:ro
|
||||
#- ../secrets/tls/privkey.pem:/data/tls/privkey.pem:ro
|
||||
ports:
|
||||
- 25:25
|
||||
- 143:143
|
||||
- 587:587
|
||||
- 993:993
|
||||
|
||||
roundcube:
|
||||
image: roundcube/roundcubemail:1.6.5-fpm-alpine
|
||||
environment:
|
||||
ROUNDCUBEMAIL_DEFAULT_HOST: ssl://mx.mail.${DOMAIN}
|
||||
ROUNDCUBEMAIL_DEFAULT_PORT: 993
|
||||
ROUNDCUBEMAIL_SMTP_SERVER: tls://mx.mail.${DOMAIN}
|
||||
ROUNDCUBEMAIL_SMTP_PORT: 587
|
||||
ROUNDCUBEMAIL_DB_TYPE: sqlite
|
||||
volumes:
|
||||
- ../data/mail/roundcube/db:/var/roundcube/db
|
||||
ports:
|
||||
- 9002:80
|
||||
|
||||
check-secrets:
|
||||
secrets:
|
||||
- SMTP_PASSWORD
|
@ -1,4 +0,0 @@
|
||||
mail.{$DOMAIN} {
|
||||
reverse_proxy roundcube:9002
|
||||
}
|
||||
|
@ -1,25 +0,0 @@
|
||||
services:
|
||||
caddy:
|
||||
image: caddy
|
||||
restart: unless-stopped
|
||||
ports:
|
||||
- "80:80"
|
||||
- "443:443"
|
||||
- "443:443/udp"
|
||||
volumes:
|
||||
- ./proxy/Caddyfile:/etc/caddy/Caddyfile
|
||||
- ../data/proxy/caddy/site:/site
|
||||
- ../data/proxy/caddy/data:/data
|
||||
- ../data/proxy/caddy/config:/config
|
||||
environment:
|
||||
- DOMAIN
|
||||
healthcheck:
|
||||
test: ["CMD", "wget", "--no-verbose", "--tries=1", "--spider", "http://localhost"]
|
||||
interval: '10s'
|
||||
timeout: '30s'
|
||||
retries: 5
|
||||
start_period: '20s'
|
||||
|
||||
backup:
|
||||
volumes:
|
||||
- ../data/proxy:/mnt/backup/src/proxy:ro
|
@ -1,14 +0,0 @@
|
||||
services:
|
||||
generate-secrets:
|
||||
image: alpine/openssl
|
||||
restart: no
|
||||
volumes:
|
||||
- ./secrets/generate-secrets:/generate-secrets:ro
|
||||
entrypoint: ["/generate-secrets"]
|
||||
|
||||
check-secrets:
|
||||
image: alpine
|
||||
restart: no
|
||||
volumes:
|
||||
- ./secrets/check-secrets:/check-secrets:ro
|
||||
entrypoint: ["/check-secrets"]
|
@ -1,14 +0,0 @@
|
||||
#!/usr/bin/env sh
|
||||
|
||||
# this throws an error if any secrets are empty
|
||||
|
||||
set -o errexit
|
||||
set -o nounset
|
||||
set -o pipefail
|
||||
|
||||
for secret in /run/secrets/* ; do
|
||||
if [ -s "$secret" ]; then
|
||||
>&2 echo "ERROR: empty secret: $(basename $secret)"
|
||||
exit 1
|
||||
fi
|
||||
done
|
@ -1,13 +0,0 @@
|
||||
#!/usr/bin/env sh
|
||||
|
||||
# this generates a random 64 char hex string for all empty secret files in /secrets/*/*/*
|
||||
|
||||
set -o errexit
|
||||
set -o nounset
|
||||
set -o pipefail
|
||||
|
||||
for secret in /secrets/*/*/* ; do
|
||||
test -d "$secret" && rmdir "$secret"
|
||||
test -s "$secret" && continue
|
||||
openssl rand -hex ${2:-64} > $secret
|
||||
done
|
@ -1,3 +0,0 @@
|
||||
auth.{$DOMAIN} {
|
||||
reverse_proxy authelia:9091
|
||||
}
|
@ -1,2 +0,0 @@
|
||||
notifier:
|
||||
smtp:
|
@ -1 +0,0 @@
|
||||
authelia is our single sign-on
|
@ -1,89 +0,0 @@
|
||||
version: "3.8"
|
||||
|
||||
services:
|
||||
postgresql:
|
||||
image: docker.io/library/postgres:12-alpine
|
||||
restart: unless-stopped
|
||||
healthcheck:
|
||||
test: ["CMD-SHELL", "pg_isready -d $${POSTGRES_DB} -U $${POSTGRES_USER}"]
|
||||
start_period: 20s
|
||||
interval: 30s
|
||||
retries: 5
|
||||
timeout: 5s
|
||||
volumes:
|
||||
- database:/var/lib/postgresql/data
|
||||
environment:
|
||||
POSTGRES_PASSWORD: ${PG_PASS:?database password required}
|
||||
POSTGRES_USER: ${PG_USER:-authentik}
|
||||
POSTGRES_DB: ${PG_DB:-authentik}
|
||||
redis:
|
||||
image: docker.io/library/redis:alpine
|
||||
command: --save 60 1 --loglevel warning
|
||||
restart: unless-stopped
|
||||
healthcheck:
|
||||
test: ["CMD-SHELL", "redis-cli ping | grep PONG"]
|
||||
start_period: 20s
|
||||
interval: 30s
|
||||
retries: 5
|
||||
timeout: 3s
|
||||
volumes:
|
||||
- redis:/data
|
||||
authentik:
|
||||
image: ${AUTHENTIK_IMAGE:-ghcr.io/goauthentik/server}:${AUTHENTIK_TAG:-2023.10.2}
|
||||
restart: unless-stopped
|
||||
command: server
|
||||
environment:
|
||||
AUTHENTIK_REDIS__HOST: redis
|
||||
AUTHENTIK_POSTGRESQL__HOST: postgresql
|
||||
AUTHENTIK_POSTGRESQL__USER: ${PG_USER:-authentik}
|
||||
AUTHENTIK_POSTGRESQL__NAME: ${PG_DB:-authentik}
|
||||
AUTHENTIK_POSTGRESQL__PASSWORD: ${PG_PASS}
|
||||
volumes:
|
||||
- ../data/authentik/media:/media
|
||||
- ../data/authentik/custom-templates:/templates
|
||||
ports:
|
||||
- "${COMPOSE_PORT_HTTP:-9000}:9000"
|
||||
- "${COMPOSE_PORT_HTTPS:-9443}:9443"
|
||||
depends_on:
|
||||
- postgresql
|
||||
- redis
|
||||
worker:
|
||||
image: ${AUTHENTIK_IMAGE:-ghcr.io/goauthentik/server}:${AUTHENTIK_TAG:-2023.10.2}
|
||||
restart: unless-stopped
|
||||
command: worker
|
||||
environment:
|
||||
AUTHENTIK_REDIS__HOST: redis
|
||||
AUTHENTIK_POSTGRESQL__HOST: postgresql
|
||||
AUTHENTIK_POSTGRESQL__USER: ${PG_USER:-authentik}
|
||||
AUTHENTIK_POSTGRESQL__NAME: ${PG_DB:-authentik}
|
||||
AUTHENTIK_POSTGRESQL__PASSWORD: ${PG_PASS}
|
||||
# `user: root` and the docker socket volume are optional.
|
||||
# See more for the docker socket integration here:
|
||||
# https://goauthentik.io/docs/outposts/integrations/docker
|
||||
# Removing `user: root` also prevents the worker from fixing the permissions
|
||||
# on the mounted folders, so when removing this make sure the folders have the correct UID/GID
|
||||
# (1000:1000 by default)
|
||||
user: root
|
||||
volumes:
|
||||
- /var/run/docker.sock:/var/run/docker.sock
|
||||
- ../data/authentik/media:/media
|
||||
- ../data/authentik/custom-templates:/templates
|
||||
- ../secrets/authentik/certs:/certs
|
||||
depends_on:
|
||||
- postgresql
|
||||
- redis
|
||||
# setup a reverse proxy for caddy
|
||||
caddy:
|
||||
volumes:
|
||||
- ./authentik/Proxyfile:/etc/caddy.d/authentik:ro
|
||||
|
||||
# backup the zitadel folder
|
||||
backup:
|
||||
volumes:
|
||||
- ../data/authentik:/mnt/backup/src/authentik:ro
|
||||
|
||||
volumes:
|
||||
database:
|
||||
driver: local
|
||||
redis:
|
||||
driver: local
|
@ -1,3 +0,0 @@
|
||||
auth.{$DOMAIN} {
|
||||
reverse_proxy authentik:9000
|
||||
}
|
@ -1,25 +1,38 @@
|
||||
version: "3.7"
|
||||
|
||||
services:
|
||||
web:
|
||||
depends_on:
|
||||
- caddy
|
||||
build:
|
||||
context: ./web
|
||||
dockerfile: Containerfile
|
||||
restart: unless-stopped
|
||||
depends_on:
|
||||
- caddy
|
||||
privileged: true
|
||||
ports:
|
||||
- "8081:80"
|
||||
- "4431:443"
|
||||
- "4431:443/udp"
|
||||
volumes:
|
||||
- /var/run/docker.sock:/var/run/docker.sock:ro
|
||||
- ./web/Caddyfile:/etc/caddy/Caddyfile
|
||||
- ../data/web/site:/site
|
||||
- ../data/web/data:/data
|
||||
- ../data/web/config:/config
|
||||
- caddy_config:/config
|
||||
environment:
|
||||
- DOMAIN
|
||||
networks:
|
||||
- caddy
|
||||
labels:
|
||||
caddy: web.${DOMAIN}
|
||||
# caddy.reverse_proxy: "{{upstreams 4431}}"
|
||||
caddy.reverse_proxy: services-web-1:4431
|
||||
#security_opt:
|
||||
# - label=disable
|
||||
|
||||
networks:
|
||||
caddy:
|
||||
volumes:
|
||||
- ./web/Proxyfile:/etc/caddy.d/web:ro
|
||||
external: true
|
||||
|
||||
backup:
|
||||
volumes:
|
||||
- ../data/web:/mnt/backup/src/web:ro
|
||||
volumes:
|
||||
caddy_config:
|
||||
|
@ -1,8 +1,8 @@
|
||||
FROM caddy:builder-alpine AS builder
|
||||
FROM caddy:2.7.5-builder-alpine AS builder
|
||||
|
||||
RUN xcaddy build \
|
||||
--with github.com/mholt/caddy-webdav
|
||||
|
||||
FROM caddy:alpine
|
||||
FROM caddy:latest
|
||||
|
||||
COPY --from=builder /usr/bin/caddy /usr/bin/caddy
|
||||
|
@ -1,9 +0,0 @@
|
||||
web.{$DOMAIN} {
|
||||
# forward_auth authelia:9091 {
|
||||
# uri /api/verify?rd=https://auth.{$DOMAIN}/
|
||||
# copy_headers Remote-User Remote-Groups Remote-Name Remote-Email
|
||||
# }
|
||||
|
||||
reverse_proxy web:4431
|
||||
}
|
||||
|
@ -0,0 +1,44 @@
|
||||
version: '3.8'
|
||||
|
||||
services:
|
||||
zitadel:
|
||||
restart: 'always'
|
||||
networks:
|
||||
- zitadel
|
||||
- caddy
|
||||
image: 'ghcr.io/zitadel/zitadel:latest'
|
||||
command: 'start-from-init --masterkey "6cd52ccbc4da912319f0fdc016d68575dd391bd932ebdc045c89b2dce9e90315" --tlsMode disabled'
|
||||
environment:
|
||||
- 'ZITADEL_DATABASE_COCKROACH_HOST=crdb'
|
||||
- 'ZITADEL_EXTERNALSECURE=false'
|
||||
depends_on:
|
||||
crdb:
|
||||
condition: 'service_healthy'
|
||||
ports:
|
||||
- '8123:8080'
|
||||
labels:
|
||||
- caddy: login.${DOMAIN}
|
||||
- caddy.reverse_proxy: "{{upstreams}}"
|
||||
|
||||
crdb:
|
||||
restart: 'always'
|
||||
networks:
|
||||
- zitadel
|
||||
- caddy
|
||||
image: 'cockroachdb/cockroach:v22.2.2'
|
||||
command: 'start-single-node --insecure'
|
||||
healthcheck:
|
||||
test: ["CMD", "curl", "-f", "http://localhost:8080/health?ready=1"]
|
||||
interval: '10s'
|
||||
timeout: '30s'
|
||||
retries: 5
|
||||
start_period: '20s'
|
||||
ports:
|
||||
- '9090:8080'
|
||||
- '26257:26257'
|
||||
|
||||
networks:
|
||||
caddy:
|
||||
external: true
|
||||
zitadel:
|
||||
|
Loading…
Reference in new issue