Compare commits
21 Commits
cb3d12484b
...
0b7f731764
| Author | SHA1 | Date | |
|---|---|---|---|
| 0b7f731764 | |||
| 733f5e2504 | |||
| 2e2211b26e | |||
| ece058e9da | |||
| 6e157cf8ee | |||
| bf4e850655 | |||
| 981da971cd | |||
| 8fc3c0ae37 | |||
| 2289f7f400 | |||
| 5feac4aa32 | |||
| a81f66fa6b | |||
| 8320490b21 | |||
| a2d8f10c2a | |||
| d1677f92c6 | |||
| 300ad5b49c | |||
| 7fbeb3aa26 | |||
| 31921e5a40 | |||
| 14c0a4f522 | |||
| 12773ef7e7 | |||
| c252250db0 | |||
| 39d5ad4b52 |
@@ -4,6 +4,7 @@ on:
|
|||||||
push:
|
push:
|
||||||
branches:
|
branches:
|
||||||
- main
|
- main
|
||||||
|
- develop
|
||||||
pull_request:
|
pull_request:
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
@@ -30,6 +31,5 @@ jobs:
|
|||||||
file: ./Dockerfile
|
file: ./Dockerfile
|
||||||
push: true
|
push: true
|
||||||
tags: |
|
tags: |
|
||||||
gitea.calahilstudios.com/${{ github.repository_owner }}/${{ github.event.repository.name }}:latest
|
gitea.calahilstudios.com/${{ github.repository_owner }}/${{ github.event.repository.name }}:develop
|
||||||
gitea.calahilstudios.com/${{ github.repository_owner }}/${{ github.event.repository.name }}:${{ github.sha }}
|
gitea.calahilstudios.com/${{ github.repository_owner }}/${{ github.event.repository.name }}:${{ github.sha }}
|
||||||
|
|
||||||
|
|||||||
1
.gitignore
vendored
1
.gitignore
vendored
@@ -0,0 +1 @@
|
|||||||
|
.env
|
||||||
|
|||||||
57
Dockerfile
57
Dockerfile
@@ -1,37 +1,50 @@
|
|||||||
FROM ghcr.io/linuxserver/duplicati:2.1.0
|
# Use LinuxServer.io Duplicati base
|
||||||
|
FROM linuxserver/duplicati:2.1.0
|
||||||
|
|
||||||
ENV DEBIAN_FRONTEND=noninteractive
|
# Install Docker CLI, bash, python3, btrfs support and all the app directories
|
||||||
SHELL ["/bin/bash", "-o", "pipefail", "-c"]
|
RUN apt-get update \
|
||||||
|
|
||||||
RUN apt-get update -y \
|
|
||||||
&& apt-get install -y --no-install-recommends \
|
&& apt-get install -y --no-install-recommends \
|
||||||
ca-certificates \
|
ca-certificates \
|
||||||
curl \
|
curl \
|
||||||
gnupg \
|
gnupg \
|
||||||
lsb-release \
|
lsb-release \
|
||||||
btrfs-progs \
|
|
||||||
#&& rm -rf /var/lib/apt/lists/* \
|
|
||||||
&& install -m 0755 -d /etc/apt/keyrings \
|
|
||||||
&& curl -fsSL https://download.docker.com/linux/ubuntu/gpg -o /etc/apt/keyrings/docker.asc \
|
|
||||||
&& chmod a+r /etc/apt/keyrings/docker.asc \
|
|
||||||
&& echo \
|
|
||||||
"deb [arch=$(dpkg --print-architecture) signed-by=/etc/apt/keyrings/docker.asc] https://download.docker.com/linux/ubuntu \
|
|
||||||
$(. /etc/os-release && echo "${UBUNTU_CODENAME:-$VERSION_CODENAME}") stable" | \
|
|
||||||
tee /etc/apt/sources.list.d/docker.list > /dev/null \
|
|
||||||
&& apt-get update -y \
|
|
||||||
&& apt-get install -y --no-install-recommends \
|
|
||||||
cron \
|
|
||||||
bash \
|
bash \
|
||||||
|
python3 \
|
||||||
|
python3-pip \
|
||||||
|
btrfs-progs \
|
||||||
|
&& mkdir -p /etc/apt/keyrings \
|
||||||
|
&& curl -fsSL "https://download.docker.com/linux/$(. /etc/os-release; echo "$ID")/gpg" \
|
||||||
|
| gpg --dearmor -o /etc/apt/keyrings/docker.gpg \
|
||||||
|
&& echo "deb [arch=$(dpkg --print-architecture) signed-by=/etc/apt/keyrings/docker.gpg] \
|
||||||
|
https://download.docker.com/linux/$(. /etc/os-release; echo "$ID") \
|
||||||
|
$(lsb_release -cs) stable" \
|
||||||
|
| tee /etc/apt/sources.list.d/docker.list > /dev/null \
|
||||||
|
&& apt-get update \
|
||||||
|
&& apt-get install -y --no-install-recommends \
|
||||||
docker-ce-cli \
|
docker-ce-cli \
|
||||||
postgresql-client \
|
&& groupadd -f docker \
|
||||||
|
&& usermod -aG docker abc \
|
||||||
&& rm -rf /var/lib/apt/lists/* \
|
&& rm -rf /var/lib/apt/lists/* \
|
||||||
&& mkdir -p /backups
|
&& mkdir -p /usr/local/bin /config /etc/services.d/backupbot
|
||||||
|
|
||||||
# Copy backup script
|
# Copy the backup script
|
||||||
COPY backup.sh /usr/local/bin/backup.sh
|
COPY backup.sh /usr/local/bin/backup.sh
|
||||||
RUN chmod +x /usr/local/bin/backup.sh \
|
RUN chmod +x /usr/local/bin/backup.sh
|
||||||
&& mkdir -p /etc/services.d/backupbot
|
|
||||||
|
# Copy the environment variables for backupbot
|
||||||
|
COPY backupbot.conf /defaults/backupbot.conf
|
||||||
|
RUN chown www-data:www-data /defaults/backupbot.conf \
|
||||||
|
&& chmod 644 /defaults/backupbot.conf
|
||||||
|
|
||||||
|
# Copy s6 service for backupbot
|
||||||
COPY services/backupbot/run /etc/services.d/backupbot/run
|
COPY services/backupbot/run /etc/services.d/backupbot/run
|
||||||
RUN chmod +x /etc/services.d/backupbot/run
|
RUN chmod +x /etc/services.d/backupbot/run
|
||||||
|
|
||||||
|
# Copy web frontend
|
||||||
|
COPY web /app
|
||||||
|
RUN chmod +x /app/cgi-bin/backupbot.cgi
|
||||||
|
# Expose web frontend port
|
||||||
|
EXPOSE 8080
|
||||||
|
|
||||||
|
# Keep duplicati entrypoint
|
||||||
|
ENTRYPOINT ["/init"]
|
||||||
|
|||||||
17
backup.sh
17
backup.sh
@@ -4,7 +4,6 @@
|
|||||||
# Author: Calahil Studios
|
# Author: Calahil Studios
|
||||||
|
|
||||||
# === CONFIGURATION ===
|
# === CONFIGURATION ===
|
||||||
LOG_FILE="$1"
|
|
||||||
BACKUP_DIR="/backups/postgres_dumps"
|
BACKUP_DIR="/backups/postgres_dumps"
|
||||||
RETENTION_DAYS="${RETENTION_DAYS:-7}" # Keep 7 days of backups
|
RETENTION_DAYS="${RETENTION_DAYS:-7}" # Keep 7 days of backups
|
||||||
|
|
||||||
@@ -19,12 +18,12 @@ ghcr.io/immich-app/postgres:14-vectorchord0.4.3-pgvectors0.2.0
|
|||||||
EOF
|
EOF
|
||||||
)
|
)
|
||||||
|
|
||||||
echo "[BACKUPBOT_INFO] Starting PostgreSQL backup service..." | tee -a "$LOG_FILE"
|
echo "[BACKUPBOT_INFO] Starting PostgreSQL backup service..."
|
||||||
mkdir -p "$BACKUP_DIR"
|
mkdir -p "$BACKUP_DIR"
|
||||||
|
|
||||||
TIMESTAMP=$(date +'%Y-%m-%d_%H-%M-%S')
|
TIMESTAMP=$(date +'%Y-%m-%d_%H-%M-%S')
|
||||||
echo "[BACKUPBOT_INFO] $(date) - Starting backup cycle ($TIMESTAMP)" | tee -a "$LOG_FILE"
|
echo "[BACKUPBOT_INFO] $(date) - Starting backup cycle ($TIMESTAMP)"
|
||||||
echo "[BACKUPBOT_INFO] Checking for running Postgres containers..." | tee -a "$LOG_FILE"
|
echo "[BACKUPBOT_INFO] Checking for running Postgres containers..."
|
||||||
|
|
||||||
# Find running containers matching known image names
|
# Find running containers matching known image names
|
||||||
MATCHING_CONTAINERS=$(
|
MATCHING_CONTAINERS=$(
|
||||||
@@ -41,7 +40,7 @@ MATCHING_CONTAINERS=$(
|
|||||||
)
|
)
|
||||||
|
|
||||||
if [ -z "$MATCHING_CONTAINERS" ]; then
|
if [ -z "$MATCHING_CONTAINERS" ]; then
|
||||||
echo "[BACKUPBOT_WARN] No Postgres containers found." | tee -a "$LOG_FILE"
|
echo "[BACKUPBOT_WARN] No Postgres containers found."
|
||||||
else
|
else
|
||||||
for container in $MATCHING_CONTAINERS; do
|
for container in $MATCHING_CONTAINERS; do
|
||||||
NAME=$(docker inspect --format '{{.Name}}' "$container" | sed 's#^/##')
|
NAME=$(docker inspect --format '{{.Name}}' "$container" | sed 's#^/##')
|
||||||
@@ -54,16 +53,16 @@ else
|
|||||||
PG_USER=$(docker inspect --format '{{range .Config.Env}}{{println .}}{{end}}' "$container" | grep POSTGRES_USER | cut -d= -f2)
|
PG_USER=$(docker inspect --format '{{range .Config.Env}}{{println .}}{{end}}' "$container" | grep POSTGRES_USER | cut -d= -f2)
|
||||||
PG_PASS=$(docker inspect --format '{{range .Config.Env}}{{println .}}{{end}}' "$container" | grep POSTGRES_PASSWORD | cut -d= -f2)
|
PG_PASS=$(docker inspect --format '{{range .Config.Env}}{{println .}}{{end}}' "$container" | grep POSTGRES_PASSWORD | cut -d= -f2)
|
||||||
if docker exec -e PGPASSWORD="$PG_PASS" "$container" pg_dumpall -U "$PG_USER" -h 127.0.0.1 >"$FILE" 2>/tmp/pg_backup_error.log; then
|
if docker exec -e PGPASSWORD="$PG_PASS" "$container" pg_dumpall -U "$PG_USER" -h 127.0.0.1 >"$FILE" 2>/tmp/pg_backup_error.log; then
|
||||||
echo "[BACKUPBOT_SUCCESS] Backup complete for $NAME -> $FILE" | tee -a "$LOG_FILE"
|
echo "[BACKUPBOT_SUCCESS] Backup complete for $NAME -> $FILE"
|
||||||
else
|
else
|
||||||
echo "[BACKUPBOT_ERROR] Backup failed for $NAME (check /tmp/pg_backup_error.log)" | tee -a "$LOG_FILE"
|
echo "[BACKUPBOT_ERROR] Backup failed for $NAME (check /tmp/pg_backup_error.log)"
|
||||||
fi
|
fi
|
||||||
# Retention cleanup
|
# Retention cleanup
|
||||||
find "$CONTAINER_BACKUP_DIR" -type f -mtime +$RETENTION_DAYS -name '*.sql' -delete
|
find "$CONTAINER_BACKUP_DIR" -type f -mtime +$RETENTION_DAYS -name '*.sql' -delete
|
||||||
done
|
done
|
||||||
fi
|
fi
|
||||||
|
|
||||||
echo "[BACKUPBOT_INFO] Creating a snapshot of /srv/appdata" | tee -a "$LOG_FILE"
|
echo "[BACKUPBOT_INFO] Creating a snapshot of /srv/appdata"
|
||||||
btrfs subvolume snapshot -r /source/appdata /backups/snapshots/$(hostname)-$(date +%F)
|
btrfs subvolume snapshot -r /source/appdata /backups/snapshots/$(hostname)-$(date +%F)
|
||||||
|
|
||||||
echo "[BACKUPBOT_INFO] Backup cycle complete." | tee -a "$LOG_FILE"
|
echo "[BACKUPBOT_INFO] Backup cycle complete."
|
||||||
|
|||||||
9
backupbot.conf
Normal file
9
backupbot.conf
Normal file
@@ -0,0 +1,9 @@
|
|||||||
|
TZ=America/Los_Angeles
|
||||||
|
BACKUP_DIR=/backups/postgres
|
||||||
|
LOG_FILE=/config/log/pgbackup.log
|
||||||
|
MAX_RETRIES=3
|
||||||
|
GOTIFY_URL=http://gotify.example.com
|
||||||
|
GOTIFY_TOKEN=your_gotify_token_here
|
||||||
|
BACKUP_HOUR=03
|
||||||
|
BACKUP_MINUTE=00
|
||||||
|
BACKUPBOT_WEB_LOGGING=DEBUG
|
||||||
@@ -1,11 +1,11 @@
|
|||||||
services:
|
services:
|
||||||
backupbot:
|
backupbot:
|
||||||
image: gitea.calahilstudios.com/calahil/backupbot:latest
|
build: .
|
||||||
container_name: backupbot
|
container_name: backupbot
|
||||||
privileged: true
|
privileged: true
|
||||||
environment:
|
environment:
|
||||||
- PUID=0
|
- PUID=1000
|
||||||
- PGID=0
|
- PGID=1000
|
||||||
- TZ=Etc/UTC
|
- TZ=Etc/UTC
|
||||||
- SETTINGS_ENCRYPTION_KEY=${KEY}
|
- SETTINGS_ENCRYPTION_KEY=${KEY}
|
||||||
- CLI_ARGS= #optional
|
- CLI_ARGS= #optional
|
||||||
@@ -14,12 +14,11 @@ services:
|
|||||||
# Config dir for duplicati
|
# Config dir for duplicati
|
||||||
- /srv/appdata/duplicati/config:/config
|
- /srv/appdata/duplicati/config:/config
|
||||||
# Backup folder to store dumps/backups
|
# Backup folder to store dumps/backups
|
||||||
- /srv/backups:/backups
|
- /srv/backups:/backups:rshared
|
||||||
# Local docker config dirs
|
|
||||||
- /srv/appdata:/source/appdata:rshared
|
|
||||||
# Docker socket to list containers
|
# Docker socket to list containers
|
||||||
- /var/run/docker.sock:/var/run/docker.sock:ro
|
- /var/run/docker.sock:/var/run/docker.sock:ro
|
||||||
|
|
||||||
ports:
|
ports:
|
||||||
- 8200:8200
|
- 8200:8200
|
||||||
|
- 8201:8080
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
|
|||||||
@@ -1,27 +1,79 @@
|
|||||||
#!/usr/bin/with-contenv bash
|
#!/usr/bin/with-contenv bash
|
||||||
set -e
|
set -e
|
||||||
|
# Source env if available
|
||||||
|
if [[ -f /config/backupbot.conf ]]; then
|
||||||
|
set -a
|
||||||
|
source /config/backupbot.conf
|
||||||
|
set +a
|
||||||
|
else
|
||||||
|
echo "[INFO] copying config vars from defaults..."
|
||||||
|
cp -r /defaults/backupbot.conf /config/
|
||||||
|
set -a
|
||||||
|
source /config/backupbot.conf
|
||||||
|
set +a
|
||||||
|
fi
|
||||||
|
|
||||||
echo "[BACKUPBOT_INFO] Starting PostgreSQL backup loop service..."
|
# Start Python HTTP server for web config in background
|
||||||
|
cd /app
|
||||||
|
|
||||||
INTERVAL_HOURS="${INTERVAL_HOURS:-24}"
|
nohup python3 -m http.server 8080 --cgi 2>&1 &
|
||||||
|
|
||||||
|
# Start backup scheduler
|
||||||
STATE_FILE="/config/last_backup_date"
|
STATE_FILE="/config/last_backup_date"
|
||||||
LOG_FILE="/config/log/pgbackup.log"
|
|
||||||
|
|
||||||
mkdir -p "$(dirname "$STATE_FILE")" "$(dirname "$LOG_FILE")"
|
# TZ
|
||||||
|
: "${TZ:=UTC}"
|
||||||
|
export TZ
|
||||||
|
|
||||||
|
# Retry config
|
||||||
|
RETRIES=3
|
||||||
|
GOTIFY_URL="${GOTIFY_URL:-}"
|
||||||
|
GOTIFY_TOKEN="${GOTIFY_TOKEN:-}"
|
||||||
|
|
||||||
|
# Helper: seconds until next 3AM
|
||||||
|
seconds_until_next_3am() {
|
||||||
|
local now next_3am
|
||||||
|
now=$(date +%s)
|
||||||
|
next_3am=$(date -d "today 03:00" +%s)
|
||||||
|
((now >= next_3am)) && next_3am=$(date -d "tomorrow 03:00" +%s)
|
||||||
|
echo $((next_3am - now))
|
||||||
|
}
|
||||||
|
|
||||||
|
# Run backup with retries
|
||||||
|
run_backup() {
|
||||||
|
local attempt=1
|
||||||
|
while ((attempt <= RETRIES)); do
|
||||||
|
echo "[INFO] Backup attempt $attempt"
|
||||||
|
if /usr/local/bin/backup.sh; then
|
||||||
|
echo "[SUCCESS] Backup completed"
|
||||||
|
return 0
|
||||||
|
else
|
||||||
|
echo "[WARN] Backup failed on attempt $attempt"
|
||||||
|
((attempt++))
|
||||||
|
sleep 5
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
# Send Gotify notification if configured
|
||||||
|
if [[ -n "$GOTIFY_URL" && -n "$GOTIFY_TOKEN" ]]; then
|
||||||
|
curl -s -X POST "$GOTIFY_URL/message?token=$GOTIFY_TOKEN" \
|
||||||
|
-F "title=Backup Failed" \
|
||||||
|
-F "message=PostgreSQL backup failed after $RETRIES attempts" \
|
||||||
|
-F "priority=5"
|
||||||
|
fi
|
||||||
|
return 1
|
||||||
|
}
|
||||||
|
|
||||||
|
# Main loop
|
||||||
while true; do
|
while true; do
|
||||||
TODAY=$(date +%F)
|
TODAY=$(date +%F)
|
||||||
|
|
||||||
# Check if a backup already ran today
|
|
||||||
if [[ -f "$STATE_FILE" && "$(cat "$STATE_FILE")" == "$TODAY" ]]; then
|
if [[ -f "$STATE_FILE" && "$(cat "$STATE_FILE")" == "$TODAY" ]]; then
|
||||||
echo "[BACKUPBOT_INFO] Backup already completed today ($TODAY). Skipping."
|
echo "[INFO] Backup already done for $TODAY"
|
||||||
else
|
else
|
||||||
echo "[BACKUPBOT_INFO] Triggering backup.sh at $(date)"
|
echo "[INFO] Running backup for $TODAY"
|
||||||
/usr/local/bin/backup.sh "$LOG_FILE"
|
if run_backup; then
|
||||||
echo "$TODAY" >"$STATE_FILE"
|
echo "$TODAY" >"$STATE_FILE"
|
||||||
echo "[BACKUPBOT_INFO] Backup completed and date recorded."
|
fi
|
||||||
fi
|
fi
|
||||||
|
SECONDS_TO_WAIT=$(seconds_until_next_3am)
|
||||||
echo "[BACKUPBOT_INFO] Sleeping for $INTERVAL_HOURS hours..."
|
sleep "$SECONDS_TO_WAIT"
|
||||||
sleep "${INTERVAL_HOURS}h"
|
|
||||||
done
|
done
|
||||||
|
|||||||
115
web/cgi-bin/backupbot.cgi
Normal file
115
web/cgi-bin/backupbot.cgi
Normal file
@@ -0,0 +1,115 @@
|
|||||||
|
#!/usr/bin/env python3
|
||||||
|
import cgi
|
||||||
|
import cgitb
|
||||||
|
import os
|
||||||
|
import json
|
||||||
|
import sys
|
||||||
|
import traceback
|
||||||
|
import tempfile
|
||||||
|
|
||||||
|
cgitb.enable()
|
||||||
|
print("Content-Type: application/json\n")
|
||||||
|
|
||||||
|
ENV_FILE = "/config/backupbot.conf"
|
||||||
|
ZONEINFO_DIR = "/usr/share/zoneinfo"
|
||||||
|
|
||||||
|
# Logging level from environment
|
||||||
|
LOG_LEVEL = os.environ.get("BACKUPBOT_WEB_LOGGING", "info").lower()
|
||||||
|
LOG_LEVELS = {"debug": 3, "info": 2, "warn": 1}
|
||||||
|
|
||||||
|
|
||||||
|
def log(level, message, exc=None):
|
||||||
|
"""
|
||||||
|
Docker-friendly logging.
|
||||||
|
level: "debug", "info", "warn"
|
||||||
|
exc: exception object (only used in debug)
|
||||||
|
"""
|
||||||
|
if LOG_LEVELS.get(level, 0) <= LOG_LEVELS.get(LOG_LEVEL, 0):
|
||||||
|
timestamp = (
|
||||||
|
__import__("datetime")
|
||||||
|
.datetime.now()
|
||||||
|
.strftime(
|
||||||
|
"%Y-%m-%d \
|
||||||
|
%H:%M:%S"
|
||||||
|
)
|
||||||
|
)
|
||||||
|
msg = f"[{timestamp}] [{level.upper()}] {message}"
|
||||||
|
print(msg, file=sys.stderr)
|
||||||
|
if exc and LOG_LEVEL == "debug":
|
||||||
|
traceback.print_exception(
|
||||||
|
type(exc), exc, exc.__traceback__, file=sys.stderr
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def read_env():
|
||||||
|
env = {}
|
||||||
|
if os.path.exists(ENV_FILE):
|
||||||
|
try:
|
||||||
|
with open(ENV_FILE) as f:
|
||||||
|
for line in f:
|
||||||
|
line = line.strip()
|
||||||
|
if not line or "=" not in line:
|
||||||
|
continue
|
||||||
|
key, val = line.split("=", 1)
|
||||||
|
env[key.strip()] = val.strip()
|
||||||
|
except Exception as e:
|
||||||
|
log("warn", f"Failed to read config: {e}", e)
|
||||||
|
return env
|
||||||
|
|
||||||
|
|
||||||
|
def write_env(env):
|
||||||
|
try:
|
||||||
|
dir_name = os.path.dirname(ENV_FILE)
|
||||||
|
os.makedirs(dir_name, exist_ok=True)
|
||||||
|
# Write atomically to temp file
|
||||||
|
with tempfile.NamedTemporaryFile("w", dir=dir_name, delete=False) as tmp:
|
||||||
|
for key, val in env.items():
|
||||||
|
tmp.write(f"{key}={val}\n")
|
||||||
|
temp_name = tmp.name
|
||||||
|
os.replace(temp_name, ENV_FILE)
|
||||||
|
log("info", f"Configuration saved to {ENV_FILE}")
|
||||||
|
except Exception as e:
|
||||||
|
log("warn", f"Failed to write config: {e}", e)
|
||||||
|
raise
|
||||||
|
|
||||||
|
|
||||||
|
def list_timezones():
|
||||||
|
zones = []
|
||||||
|
for root, _, files in os.walk(ZONEINFO_DIR):
|
||||||
|
rel_root = os.path.relpath(root, ZONEINFO_DIR)
|
||||||
|
if rel_root.startswith(("posix", "right")):
|
||||||
|
continue
|
||||||
|
for file in files:
|
||||||
|
if file.startswith(".") or file.endswith((".tab", ".zi")):
|
||||||
|
continue
|
||||||
|
zones.append(os.path.join(rel_root, file) if rel_root != "." else file)
|
||||||
|
return sorted(zones)
|
||||||
|
|
||||||
|
|
||||||
|
form = cgi.FieldStorage()
|
||||||
|
action = form.getvalue("action")
|
||||||
|
|
||||||
|
try:
|
||||||
|
if action == "get":
|
||||||
|
env = read_env()
|
||||||
|
log("debug", f"Returning configuration: {env}")
|
||||||
|
print(json.dumps(env))
|
||||||
|
elif action == "set":
|
||||||
|
raw_len = os.environ.get("CONTENT_LENGTH")
|
||||||
|
length = int(raw_len) if raw_len else 0
|
||||||
|
data = json.loads(os.read(0, length))
|
||||||
|
log("debug", f"Received new configuration: {data}")
|
||||||
|
env = read_env()
|
||||||
|
env.update(data) # update existing keys, add new keys
|
||||||
|
write_env(env)
|
||||||
|
print(json.dumps({"status": "ok", "message": "Configuration saved."}))
|
||||||
|
elif action == "get_timezones":
|
||||||
|
zones = list_timezones()
|
||||||
|
log("debug", f"Returning {len(zones)} timezones")
|
||||||
|
print(json.dumps({"timezones": zones}))
|
||||||
|
else:
|
||||||
|
log("warn", f"Invalid action requested: {action}")
|
||||||
|
print(json.dumps({"status": "error", "message": "Invalid action"}))
|
||||||
|
except Exception as e:
|
||||||
|
log("warn", f"Unhandled exception: {e}", e)
|
||||||
|
print(json.dumps({"status": "error", "message": str(e)}))
|
||||||
118
web/index.html
Normal file
118
web/index.html
Normal file
@@ -0,0 +1,118 @@
|
|||||||
|
<!DOCTYPE html>
|
||||||
|
<html lang="en">
|
||||||
|
|
||||||
|
<head>
|
||||||
|
<meta charset="UTF-8">
|
||||||
|
<title>BackupBot Configuration</title>
|
||||||
|
<style>
|
||||||
|
body {
|
||||||
|
font-family: sans-serif;
|
||||||
|
margin: 2rem;
|
||||||
|
background: #f4f4f4;
|
||||||
|
}
|
||||||
|
|
||||||
|
label {
|
||||||
|
display: block;
|
||||||
|
margin-top: 1rem;
|
||||||
|
}
|
||||||
|
|
||||||
|
input {
|
||||||
|
width: 200px;
|
||||||
|
}
|
||||||
|
|
||||||
|
button {
|
||||||
|
margin-top: 1rem;
|
||||||
|
padding: 0.5rem 1rem;
|
||||||
|
}
|
||||||
|
</style>
|
||||||
|
</head>
|
||||||
|
|
||||||
|
<body>
|
||||||
|
<h1>BackupBot Configuration</h1>
|
||||||
|
|
||||||
|
<form id="configForm">
|
||||||
|
<label>Timezone:
|
||||||
|
<select id="tzSelect" name="TZ">
|
||||||
|
<option value="">Loading...</option>
|
||||||
|
</select>
|
||||||
|
</label>
|
||||||
|
<label>Backup Directory:
|
||||||
|
<input type="text" name="BACKUP_DIR" id="backupDir" placeholder="/backups">
|
||||||
|
<button type="button" onclick="chooseBackupDir()">Browse</button>
|
||||||
|
</label>
|
||||||
|
<label>Log File:
|
||||||
|
<input type="text" name="LOG_FILE" id="logDir" placeholder="/config/log">
|
||||||
|
<button type="button" onclick="chooseLogDir()">Browse</button>
|
||||||
|
</label>
|
||||||
|
<label>Backup Hour:
|
||||||
|
<input type="number" name="BACKUP_HOUR" min="0" max="23">
|
||||||
|
</label>
|
||||||
|
<label>Backup Minute:
|
||||||
|
<input type="number" name="BACKUP_MINUTE" min="0" max="59">
|
||||||
|
</label>
|
||||||
|
<label>Max Retries:
|
||||||
|
<input type="number" name="MAX_RETRIES" min="1" max="10">
|
||||||
|
</label>
|
||||||
|
<label>Gotify URL:
|
||||||
|
<input type="text" name="GOTIFY_URL">
|
||||||
|
</label>
|
||||||
|
<label>Gotify Token:
|
||||||
|
<input type="text" name="GOTIFY_TOKEN">
|
||||||
|
</label>
|
||||||
|
<button type="submit">Save Configuration</button>
|
||||||
|
</form>
|
||||||
|
|
||||||
|
<p id="status"></p>
|
||||||
|
|
||||||
|
<script>
|
||||||
|
async function loadTimezones() {
|
||||||
|
const res = await fetch('/cgi-bin/backupbot.cgi?action=get_timezones');
|
||||||
|
const data = await res.json();
|
||||||
|
const select = document.getElementById('tzSelect');
|
||||||
|
select.innerHTML = '';
|
||||||
|
data.timezones.forEach(tz => {
|
||||||
|
const opt = document.createElement('option');
|
||||||
|
opt.value = tz;
|
||||||
|
opt.textContent = tz;
|
||||||
|
select.appendChild(opt);
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
function chooseBackupDir() {
|
||||||
|
const base = prompt("Enter or confirm your backup directory path:", "/backups");
|
||||||
|
if (base) document.getElementById('backupDir').value = base;
|
||||||
|
}
|
||||||
|
|
||||||
|
function chooseLogDir() {
|
||||||
|
const base = prompt("Enter or confirm your log directory path:", "/config/log");
|
||||||
|
if (base) document.getElementById('logDir').value = base;
|
||||||
|
}
|
||||||
|
|
||||||
|
async function loadConfig() {
|
||||||
|
const res = await fetch('/cgi-bin/backupbot.cgi?action=get');
|
||||||
|
const data = await res.json();
|
||||||
|
const form = document.getElementById('configForm');
|
||||||
|
Object.keys(data).forEach(key => {
|
||||||
|
if (form.elements[key]) form.elements[key].value = data[key];
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
async function saveConfig(e) {
|
||||||
|
e.preventDefault();
|
||||||
|
const formData = new FormData(document.getElementById('configForm'));
|
||||||
|
const obj = Object.fromEntries(formData.entries());
|
||||||
|
const res = await fetch('/cgi-bin/backupbot.cgi?action=set', {
|
||||||
|
method: 'POST',
|
||||||
|
headers: {'Content-Type': 'application/json'},
|
||||||
|
body: JSON.stringify(obj)
|
||||||
|
});
|
||||||
|
const result = await res.json();
|
||||||
|
document.getElementById('status').innerText = result.message;
|
||||||
|
}
|
||||||
|
|
||||||
|
document.getElementById('configForm').addEventListener('submit', saveConfig);
|
||||||
|
loadTimezones().then(loadConfig);
|
||||||
|
</script>
|
||||||
|
</body>
|
||||||
|
|
||||||
|
</html>
|
||||||
Reference in New Issue
Block a user