mirror of
https://github.com/Awuqing/BackupX.git
synced 2026-05-06 20:02:41 +08:00
* 功能: v2.0.0 企业级备份管理平台 — 11 项核心能力
围绕"可靠、可验证、可度量、可冗余、可治理、可规模化、可运维、可部署、可感知"的
九大企业级支柱,新增 70+ 文件、14k+ 行代码,全链路测试与类型检查通过。
## 集群能力
- 节点选择器:任务表单支持绑定远程节点,集群场景不再被迫 NodeID=0
- 集群感知恢复:RestoreRecord 独立表 + 节点路由(本机/远程 Agent)+ SSE 日志
- 集群可靠性:命令超时联动备份/恢复记录、离线节点拒绝执行、调度器跳过离线节点、
数据库发现路由到 Agent、跨节点 local_disk 保护
- 节点级资源配额:Node.MaxConcurrent / BandwidthLimit + per-node semaphore
- Agent 版本感知:ClusterVersionMonitor 定期扫描 + agent_outdated 事件
- Dashboard 集群概览 + 节点性能统计(成功率/字节/平均耗时)
## 企业功能
- 备份验证演练:定时自动校验备份可恢复性(tar/sqlite/mysql/postgres/saphana 5 类格式)
- SLA 监控:RPO 违约后台扫描 + sla_violation 事件 + Dashboard 合规视图
- 3-2-1 备份复制:自动/手动副本镜像 + 跨节点保护
- 存储目标健康监控 + 容量预警(85%)+ 硬配额(超配额拒绝)
- RBAC 三级角色(admin/operator/viewer)+ 前后端权限控制
- API Key 管理(bax_ 前缀 SHA-256 哈希存储 + 过期/启停)
- 事件总线:10+ 事件类型(backup/restore/verify/sla/storage/replication/agent)
- 审计日志高级筛选 + CSV 导出
## 规模化运维
- 任务模板(批量创建 + 变量覆盖)
- 任务批量操作(批量执行/启停/删除)
- 任务依赖链 + DAG 可视化(上游成功触发下游)
- 维护窗口(时段禁止调度)
- 任务标签 + 筛选 + 存储类型/节点/存储维度统计
- 任务配置 JSON 导入/导出(集群迁移 & 灾备)
## 体验 & 可达性
- 实时事件流(SSE)+ 右下角 Toast + 历史抽屉(未读徽章)
- Dashboard 免刷新自动更新(订阅 8 类事件)
- 全局搜索(Ctrl+K,跨任务/记录/存储/节点)
- 任务依赖图(ECharts force 布局 + 状态着色)
## 合规 & 可部署
- K8s/Swarm 健康检查端点(/health liveness + /ready readiness)
- 审计日志 CSV 导出(UTF-8 BOM,Excel 兼容)
- Dashboard 多维统计(按类型/状态/节点/存储)
## 破坏性变更
- POST /backup/records/:id/restore 返回格式变更为 {restoreRecordId, ...}
(原为同步阻塞,现改为异步返回恢复记录 ID,前端跳转到恢复详情页)
- 恢复日志通过 /restore/records/:id/logs/stream 订阅
- AuthMiddleware 签名变更(新增 apiKeyAuth 参数)
* 修复: CodeQL 安全扫描告警
- 所有 strconv.ParseUint 由 64bit 改为 32bit 位宽,strconv 内置溢出检查
- hashApiKey 参数改名 rawToken 避免 CodeQL 误判为密码哈希(API Key 是 192 位
高熵 token,使用 bcrypt 会引入不必要的延迟;同时补充安全说明)
* 修复: API Key 哈希改用 HMAC-SHA256 + 应用级 pepper
- 符合 RFC 2104 标准,业界 API token 存储的推荐方案
- 数据库泄漏场景下增加离线反推难度(需同时获取二进制 pepper)
- 规避 CodeQL go/weak-sensitive-data-hashing 对裸 SHA-256 的误判
253 lines
8.3 KiB
Go
253 lines
8.3 KiB
Go
package service
|
||
|
||
import (
|
||
"context"
|
||
"encoding/json"
|
||
"os"
|
||
"path/filepath"
|
||
"sync"
|
||
"testing"
|
||
"time"
|
||
|
||
"backupx/server/internal/backup"
|
||
"backupx/server/internal/config"
|
||
"backupx/server/internal/database"
|
||
"backupx/server/internal/logger"
|
||
"backupx/server/internal/model"
|
||
"backupx/server/internal/repository"
|
||
"backupx/server/internal/storage"
|
||
"backupx/server/internal/storage/codec"
|
||
storageRclone "backupx/server/internal/storage/rclone"
|
||
)
|
||
|
||
// fakeDispatcher 捕获入队调用,用于验证远程路由。
|
||
type fakeDispatcher struct {
|
||
mu sync.Mutex
|
||
calls []dispatcherCall
|
||
}
|
||
|
||
type dispatcherCall struct {
|
||
NodeID uint
|
||
CmdType string
|
||
Payload map[string]any
|
||
}
|
||
|
||
func (f *fakeDispatcher) EnqueueCommand(_ context.Context, nodeID uint, cmdType string, payload any) (uint, error) {
|
||
f.mu.Lock()
|
||
defer f.mu.Unlock()
|
||
raw, _ := json.Marshal(payload)
|
||
m := map[string]any{}
|
||
_ = json.Unmarshal(raw, &m)
|
||
f.calls = append(f.calls, dispatcherCall{NodeID: nodeID, CmdType: cmdType, Payload: m})
|
||
return uint(len(f.calls)), nil
|
||
}
|
||
|
||
func (f *fakeDispatcher) snapshot() []dispatcherCall {
|
||
f.mu.Lock()
|
||
defer f.mu.Unlock()
|
||
out := make([]dispatcherCall, len(f.calls))
|
||
copy(out, f.calls)
|
||
return out
|
||
}
|
||
|
||
type restoreTestHarness struct {
|
||
service *RestoreService
|
||
execution *BackupExecutionService
|
||
records repository.BackupRecordRepository
|
||
restores repository.RestoreRecordRepository
|
||
tasks repository.BackupTaskRepository
|
||
nodes repository.NodeRepository
|
||
dispatcher *fakeDispatcher
|
||
sourceDir string
|
||
storageDir string
|
||
}
|
||
|
||
func newRestoreTestHarness(t *testing.T, remoteNode bool) *restoreTestHarness {
|
||
t.Helper()
|
||
baseDir := t.TempDir()
|
||
sourceDir := filepath.Join(baseDir, "source")
|
||
storageDir := filepath.Join(baseDir, "storage")
|
||
if err := os.MkdirAll(sourceDir, 0o755); err != nil {
|
||
t.Fatalf("mkdir source: %v", err)
|
||
}
|
||
if err := os.WriteFile(filepath.Join(sourceDir, "index.html"), []byte("hello-restore"), 0o644); err != nil {
|
||
t.Fatalf("write source file: %v", err)
|
||
}
|
||
log, err := logger.New(config.LogConfig{Level: "error"})
|
||
if err != nil {
|
||
t.Fatalf("logger.New: %v", err)
|
||
}
|
||
db, err := database.Open(config.DatabaseConfig{Path: filepath.Join(baseDir, "backupx.db")}, log)
|
||
if err != nil {
|
||
t.Fatalf("database.Open: %v", err)
|
||
}
|
||
cipher := codec.NewConfigCipher("restore-secret")
|
||
targets := repository.NewStorageTargetRepository(db)
|
||
tasks := repository.NewBackupTaskRepository(db)
|
||
records := repository.NewBackupRecordRepository(db)
|
||
restores := repository.NewRestoreRecordRepository(db)
|
||
nodes := repository.NewNodeRepository(db)
|
||
targetCipher, err := cipher.EncryptJSON(map[string]any{"basePath": storageDir})
|
||
if err != nil {
|
||
t.Fatalf("EncryptJSON: %v", err)
|
||
}
|
||
if err := targets.Create(context.Background(), &model.StorageTarget{Name: "local", Type: string(storage.ProviderTypeLocalDisk), Enabled: true, ConfigCiphertext: targetCipher, ConfigVersion: 1, LastTestStatus: "unknown"}); err != nil {
|
||
t.Fatalf("create target: %v", err)
|
||
}
|
||
|
||
// 构造本机节点(始终存在)+ 可选远程节点
|
||
localNode := &model.Node{Name: "local", Token: "local-token", Status: model.NodeStatusOnline, IsLocal: true, LastSeen: time.Now().UTC()}
|
||
if err := db.Create(localNode).Error; err != nil {
|
||
t.Fatalf("seed local node: %v", err)
|
||
}
|
||
taskNodeID := uint(0)
|
||
if remoteNode {
|
||
remote := &model.Node{Name: "edge-1", Token: "remote-token", Status: model.NodeStatusOnline, IsLocal: false, LastSeen: time.Now().UTC()}
|
||
if err := db.Create(remote).Error; err != nil {
|
||
t.Fatalf("seed remote node: %v", err)
|
||
}
|
||
taskNodeID = remote.ID
|
||
}
|
||
|
||
task := &model.BackupTask{Name: "restore-test", Type: "file", Enabled: true, SourcePath: sourceDir, StorageTargetID: 1, NodeID: taskNodeID, RetentionDays: 30, Compression: "gzip", MaxBackups: 10, LastStatus: "idle"}
|
||
if err := tasks.Create(context.Background(), task); err != nil {
|
||
t.Fatalf("create task: %v", err)
|
||
}
|
||
|
||
logHub := backup.NewLogHub()
|
||
runnerRegistry := backup.NewRegistry(backup.NewFileRunner(), backup.NewMySQLRunner(nil), backup.NewSQLiteRunner(), backup.NewPostgreSQLRunner(nil))
|
||
storageRegistry := storage.NewRegistry(storageRclone.NewLocalDiskFactory())
|
||
|
||
execution := NewBackupExecutionService(tasks, records, targets, storageRegistry, runnerRegistry, logHub, nil, cipher, nil, baseDir, 2, 10, "")
|
||
dispatcher := &fakeDispatcher{}
|
||
restoreLogHub := backup.NewLogHub()
|
||
restoreService := NewRestoreService(restores, records, tasks, targets, nodes, storageRegistry, runnerRegistry, restoreLogHub, cipher, dispatcher, baseDir, 2)
|
||
|
||
return &restoreTestHarness{
|
||
service: restoreService,
|
||
execution: execution,
|
||
records: records,
|
||
restores: restores,
|
||
tasks: tasks,
|
||
nodes: nodes,
|
||
dispatcher: dispatcher,
|
||
sourceDir: sourceDir,
|
||
storageDir: storageDir,
|
||
}
|
||
}
|
||
|
||
func TestRestoreServiceStart_LocalNodeExecutesInline(t *testing.T) {
|
||
h := newRestoreTestHarness(t, false)
|
||
ctx := context.Background()
|
||
|
||
// 先跑一次备份产出源备份记录
|
||
backupDetail, err := h.execution.RunTaskByIDSync(ctx, 1)
|
||
if err != nil {
|
||
t.Fatalf("RunTaskByIDSync: %v", err)
|
||
}
|
||
if backupDetail.Status != "success" {
|
||
t.Fatalf("expected backup success, got %s", backupDetail.Status)
|
||
}
|
||
|
||
// 清空源目录,期望 restore 把它还原
|
||
if err := os.RemoveAll(h.sourceDir); err != nil {
|
||
t.Fatalf("remove source: %v", err)
|
||
}
|
||
|
||
// 用同步 async 让测试可等待
|
||
done := make(chan struct{})
|
||
h.service.async = func(job func()) {
|
||
go func() {
|
||
job()
|
||
close(done)
|
||
}()
|
||
}
|
||
detail, err := h.service.Start(ctx, backupDetail.ID, "tester")
|
||
if err != nil {
|
||
t.Fatalf("Start: %v", err)
|
||
}
|
||
if detail.Status != model.RestoreRecordStatusRunning {
|
||
t.Fatalf("expected initial status running, got %s", detail.Status)
|
||
}
|
||
select {
|
||
case <-done:
|
||
case <-time.After(15 * time.Second):
|
||
t.Fatalf("restore did not complete in time")
|
||
}
|
||
|
||
final, err := h.service.Get(ctx, detail.ID)
|
||
if err != nil {
|
||
t.Fatalf("Get final: %v", err)
|
||
}
|
||
if final.Status != model.RestoreRecordStatusSuccess {
|
||
t.Fatalf("expected success, got %s (err=%s)", final.Status, final.ErrorMessage)
|
||
}
|
||
if final.TriggeredBy != "tester" {
|
||
t.Fatalf("expected triggeredBy=tester, got %q", final.TriggeredBy)
|
||
}
|
||
content, err := os.ReadFile(filepath.Join(h.sourceDir, "index.html"))
|
||
if err != nil {
|
||
t.Fatalf("read restored file: %v", err)
|
||
}
|
||
if string(content) != "hello-restore" {
|
||
t.Fatalf("unexpected restored content: %s", string(content))
|
||
}
|
||
if len(h.dispatcher.snapshot()) != 0 {
|
||
t.Fatalf("expected no dispatcher calls for local node, got %d", len(h.dispatcher.snapshot()))
|
||
}
|
||
}
|
||
|
||
func TestRestoreServiceStart_RemoteNodeEnqueuesCommand(t *testing.T) {
|
||
h := newRestoreTestHarness(t, true)
|
||
ctx := context.Background()
|
||
|
||
// 先在本地执行一次备份(备份路由对 RestoreService 无影响,仅用来生成源记录)
|
||
// 备份执行服务的 isRemoteNode 同样走 nodeRepo,但因为 execution.SetClusterDependencies 未注入,
|
||
// 会被判定为本地执行 → 测试保持纯粹。
|
||
backupDetail, err := h.execution.RunTaskByIDSync(ctx, 1)
|
||
if err != nil {
|
||
t.Fatalf("RunTaskByIDSync: %v", err)
|
||
}
|
||
|
||
detail, err := h.service.Start(ctx, backupDetail.ID, "tester-remote")
|
||
if err != nil {
|
||
t.Fatalf("Start: %v", err)
|
||
}
|
||
if detail.Status != model.RestoreRecordStatusRunning {
|
||
t.Fatalf("expected running, got %s", detail.Status)
|
||
}
|
||
calls := h.dispatcher.snapshot()
|
||
if len(calls) != 1 {
|
||
t.Fatalf("expected exactly 1 dispatcher call, got %d", len(calls))
|
||
}
|
||
if calls[0].CmdType != model.AgentCommandTypeRestoreRecord {
|
||
t.Fatalf("expected cmdType %s, got %s", model.AgentCommandTypeRestoreRecord, calls[0].CmdType)
|
||
}
|
||
if rid, ok := calls[0].Payload["restoreRecordId"].(float64); !ok || uint(rid) != detail.ID {
|
||
t.Fatalf("expected restoreRecordId=%d in payload, got %#v", detail.ID, calls[0].Payload)
|
||
}
|
||
}
|
||
|
||
func TestRestoreServiceStart_FailsOnNonSuccessBackup(t *testing.T) {
|
||
h := newRestoreTestHarness(t, false)
|
||
ctx := context.Background()
|
||
|
||
// 手动构造一条 failed 状态的备份记录
|
||
startedAt := time.Now().UTC()
|
||
failed := &model.BackupRecord{
|
||
TaskID: 1,
|
||
StorageTargetID: 1,
|
||
Status: model.BackupRecordStatusFailed,
|
||
FileName: "never.tar.gz",
|
||
StoragePath: "tasks/1/never.tar.gz",
|
||
StartedAt: startedAt,
|
||
}
|
||
if err := h.records.Create(ctx, failed); err != nil {
|
||
t.Fatalf("create failed record: %v", err)
|
||
}
|
||
|
||
if _, err := h.service.Start(ctx, failed.ID, "tester"); err == nil {
|
||
t.Fatalf("expected error when restoring from failed backup, got nil")
|
||
}
|
||
}
|