Filecatalyst Profiles -
destination: type: "s3" bucket: "backup-bucket" path: "database/ date /" lifecycle: "delete after 90 days"
transfer: adaptive_bandwidth: true min_bandwidth: "20Mbps" max_bandwidth: "200Mbps" compression: "zstd" encryption: "AES-256-GCM" parallel_chunks: 8 verify_checksum: "SHA-256" filecatalyst profiles
filecatalyst profile test --profile new_config.yaml --dry-run filecatalyst profiles
// Exclude patterns exclude: [ " .tmp", " .partial", "/.*/" // hidden files ] Per-profile bandwidth management with priority queuing: filecatalyst profiles
notifications: on_success: "slack:#backups - Success: size transferred" on_failure: - "pagerduty:Database Backup Failed" - "email:db-admin@company.com"
# Python SDK from filecatalyst import Profile, Orchestrator