mirror of
https://github.com/Dokploy/dokploy
synced 2025-06-26 18:27:59 +00:00
- Added normalizeS3Path function to standardize S3 path formatting by trimming whitespace and removing leading/trailing slashes. - Updated backup-related modules (MySQL, MongoDB, PostgreSQL, MariaDB, and web server) to utilize normalizeS3Path for consistent S3 path handling. - Introduced unit tests for normalizeS3Path to ensure correct functionality across various input scenarios.
71 lines
2.6 KiB
TypeScript
71 lines
2.6 KiB
TypeScript
import type { BackupSchedule } from "@dokploy/server/services/backup";
|
|
import type { Postgres } from "@dokploy/server/services/postgres";
|
|
import { findProjectById } from "@dokploy/server/services/project";
|
|
import {
|
|
getRemoteServiceContainer,
|
|
getServiceContainer,
|
|
} from "../docker/utils";
|
|
import { sendDatabaseBackupNotifications } from "../notifications/database-backup";
|
|
import { execAsync, execAsyncRemote } from "../process/execAsync";
|
|
import { getS3Credentials, normalizeS3Path } from "./utils";
|
|
|
|
export const runPostgresBackup = async (
|
|
postgres: Postgres,
|
|
backup: BackupSchedule,
|
|
) => {
|
|
const { appName, databaseUser, name, projectId } = postgres;
|
|
const project = await findProjectById(projectId);
|
|
|
|
const { prefix, database } = backup;
|
|
const destination = backup.destination;
|
|
const backupFileName = `${new Date().toISOString()}.sql.gz`;
|
|
const bucketDestination = `${normalizeS3Path(prefix)}${backupFileName}`;
|
|
try {
|
|
const rcloneFlags = getS3Credentials(destination);
|
|
const rcloneDestination = `:s3:${destination.bucket}/${bucketDestination}`;
|
|
|
|
const rcloneCommand = `rclone rcat ${rcloneFlags.join(" ")} "${rcloneDestination}"`;
|
|
if (postgres.serverId) {
|
|
const { Id: containerId } = await getRemoteServiceContainer(
|
|
postgres.serverId,
|
|
appName,
|
|
);
|
|
const pgDumpCommand = `docker exec ${containerId} sh -c "pg_dump -Fc --no-acl --no-owner -h localhost -U ${databaseUser} --no-password '${database}' | gzip"`;
|
|
|
|
await execAsyncRemote(
|
|
postgres.serverId,
|
|
`${pgDumpCommand} | ${rcloneCommand}`,
|
|
);
|
|
} else {
|
|
const { Id: containerId } = await getServiceContainer(appName);
|
|
|
|
const pgDumpCommand = `docker exec ${containerId} sh -c "pg_dump -Fc --no-acl --no-owner -h localhost -U ${databaseUser} --no-password '${database}' | gzip"`;
|
|
await execAsync(`${pgDumpCommand} | ${rcloneCommand}`);
|
|
}
|
|
|
|
await sendDatabaseBackupNotifications({
|
|
applicationName: name,
|
|
projectName: project.name,
|
|
databaseType: "postgres",
|
|
type: "success",
|
|
organizationId: project.organizationId,
|
|
});
|
|
} catch (error) {
|
|
await sendDatabaseBackupNotifications({
|
|
applicationName: name,
|
|
projectName: project.name,
|
|
databaseType: "postgres",
|
|
type: "error",
|
|
// @ts-ignore
|
|
errorMessage: error?.message || "Error message not provided",
|
|
organizationId: project.organizationId,
|
|
});
|
|
|
|
throw error;
|
|
} finally {
|
|
}
|
|
};
|
|
|
|
// Restore
|
|
// /Applications/pgAdmin 4.app/Contents/SharedSupport/pg_restore --host "localhost" --port "5432" --username "mauricio" --no-password --dbname "postgres" --verbose "/Users/mauricio/Downloads/_databases_2024-04-12T07_02_05.234Z.sql"
|