الملفات
backup_tool/pages/api/backup-stream.ts

171 أسطر
6.0 KiB
TypeScript

import type { NextApiRequest, NextApiResponse } from 'next';
import { spawn } from 'child_process';
import path from 'path';
import fs from 'fs';
import { randomUUID } from 'crypto';
import { S3Client, GetObjectCommand } from "@aws-sdk/client-s3";
import { Upload } from "@aws-sdk/lib-storage";
import { getSignedUrl } from "@aws-sdk/s3-request-presigner";
import { PassThrough } from 'stream';
// --- Helper Types and Functions for History Logging ---
type BackupStatus = 'COMPLETED' | 'FAILED' | 'PROCESSING' | 'QUEUED' | 'CANCELLED';
type BackupRecord = {
id: string;
dbName: string;
status: BackupStatus;
createdAt: string;
fileName?: string;
error?: string;
downloadUrl?: string;
};
const DB_PATH = path.resolve(process.cwd(), 'backup-history.json');
const readRecords = async (): Promise<BackupRecord[]> => {
try {
if (!fs.existsSync(DB_PATH)) {
await fs.promises.writeFile(DB_PATH, JSON.stringify([]), 'utf-8');
return [];
}
const fileContent = await fs.promises.readFile(DB_PATH, 'utf-8');
return fileContent ? JSON.parse(fileContent) : [];
} catch (error) {
console.error("Error reading backup history:", error);
return [];
}
};
const writeRecords = async (records: BackupRecord[]): Promise<void> => {
try {
await fs.promises.writeFile(DB_PATH, JSON.stringify(records, null, 2), 'utf-8');
} catch (error) {
console.error("Error writing backup history:", error);
}
};
const sendStreamMessage = (res: NextApiResponse, data: object) => {
try {
if (!res.writableEnded) {
res.write(`data: ${JSON.stringify(data)}\n\n`);
}
} catch (e) {
console.error("Failed to write to stream:", e);
}
};
export default async function handler(req: NextApiRequest, res: NextApiResponse) {
res.setHeader('Content-Type', 'text/event-stream');
res.setHeader('Cache-Control', 'no-cache');
res.setHeader('Connection', 'keep-alive');
res.flushHeaders();
const {
dbHost, dbPort, dbUser, dbPassword, dbName, dbRequireSsl,
s3Endpoint, s3BucketName, s3AccessKey, s3SecretKey, s3Region
} = req.query;
const recordId = randomUUID();
const timestamp = new Date().toISOString().replace(/[:.]/g, '-');
const backupFileName = `${dbName as string}_${timestamp}.dump`;
const backupsDir = path.resolve(process.cwd(), 'backups');
const backupFilePath = path.join(backupsDir, backupFileName);
if (!fs.existsSync(backupsDir)) fs.mkdirSync(backupsDir);
const newRecord: BackupRecord = { id: recordId, dbName: dbName as string, status: 'PROCESSING', createdAt: new Date().toISOString(), fileName: backupFileName };
const records = await readRecords();
records.push(newRecord);
await writeRecords(records);
const pgDumpCommand = 'pg_dump';
const args: string[] = [
'--format=c',
'--blobs',
'--verbose',
`--host=${dbHost}`,
`--port=${dbPort}`,
`--username=${dbUser}`,
`--dbname=${dbName}`,
];
const env: NodeJS.ProcessEnv = {
...process.env,
PGPASSWORD: dbPassword as string,
PGSSLMODE: dbRequireSsl === 'true' ? 'require' : 'prefer',
};
const backupProcess = spawn(pgDumpCommand, args, { env });
req.on('close', async () => {
console.log("Client disconnected. Terminating backup process...");
backupProcess.kill(); // Stop the pg_dump process
const finalRecords = await readRecords();
const recordIndex = finalRecords.findIndex(r => r.id === recordId);
if (recordIndex > -1 && finalRecords[recordIndex].status === 'PROCESSING') {
finalRecords[recordIndex].status = 'CANCELLED';
finalRecords[recordIndex].error = 'Process cancelled by user.';
await writeRecords(finalRecords);
}
res.end();
});
const s3Client = new S3Client({
endpoint: s3Endpoint as string,
region: s3Region as string,
credentials: { accessKeyId: s3AccessKey as string, secretAccessKey: s3SecretKey as string }
});
const passThrough = new PassThrough();
passThrough.pipe(fs.createWriteStream(backupFilePath));
const s3Upload = new Upload({
client: s3Client,
params: { Bucket: s3BucketName as string, Key: backupFileName, Body: passThrough, ContentType: 'application/octet-stream' },
});
if (backupProcess.stdout) backupProcess.stdout.pipe(passThrough);
let errorOutput = '';
if (backupProcess.stderr) {
backupProcess.stderr.on('data', (data: Buffer | string) => {
errorOutput += data.toString();
sendStreamMessage(res, { message: data.toString().trim() });
});
}
try {
await s3Upload.done();
sendStreamMessage(res, { message: "✅ S3 upload completed successfully." });
const command = new GetObjectCommand({ Bucket: s3BucketName as string, Key: backupFileName });
const signedUrl = await getSignedUrl(s3Client, command, { expiresIn: 3600 });
const finalRecords = await readRecords();
const recordIndex = finalRecords.findIndex(r => r.id === recordId);
if (recordIndex > -1) {
finalRecords[recordIndex].status = 'COMPLETED';
finalRecords[recordIndex].downloadUrl = signedUrl;
await writeRecords(finalRecords);
}
sendStreamMessage(res, { message: "All tasks finished.", status: 'completed' });
} catch (err: any) {
sendStreamMessage(res, { message: `❌ S3 Upload Failed: ${err.message}`, status: 'failed' });
const finalRecords = await readRecords();
const recordIndex = finalRecords.findIndex(r => r.id === recordId);
if (recordIndex > -1) {
finalRecords[recordIndex].status = 'FAILED';
finalRecords[recordIndex].error = `S3 Error: ${err.message}`;
await writeRecords(finalRecords);
}
} finally {
if (!res.writableEnded) {
sendStreamMessage(res, { status: 'closed' });
res.end();
}
}
}