Your data is one of your most valuable assets. Unfortunately, a wide variety of threats both locally and out on the Internet seek to disrupt operations and destroy data.
Implementing a sound backup strategy is a fundamental building block to business continuity. Backing up to the cloud provides an accessible, offsite copies of data to recover from outages, disasters, human errors, malicious attacks, and more.
In this comprehensive guide, I‘ll share techniques and best practices from my years as a full-stack developer for leveraging cloud backups as a pillar of infrastructure resilience.
The Growing Threat Landscape
Businesses accumulate massive amounts of businesss-critical data across sales records, financial reports, customer information, intellectual property, and more. Losing access or control of these digital assets can inflict severe damage.
According to Veritas research, ransomware attacks grew 144% in 2022 resulting in downtime and hefty recovery costs. Hardware failures, human errors like accidental deletions, and natural disasters also contribute to data loss and business disruption.
In one study, 25% of businesses reported losing access to critical data for 24 hours every year. The average cost of downtime neared $250,000 annually, cutting productivity by 40% source.
As data volumes explode over the next 5 years, these threats will only multiply:

Image source: Enterprise Storage Forum
Without adequate precautions, companies stand little chance avoiding the crippling impact of cyber-attacks and data disasters. Now more than ever before, adopting backup best practices is an absolute necessity.
The 6-3-2 Rule for Backup Data Recovery
Industry experts at the Disaster Recovery Institute International developed the 6-3-2 rule to set standards for resilient backup implementations. This rule helps outline Recovery Time Objective (RTO) and Recovery Point Objective (RPO) targets.
**The 6-3-2 rule dictates:
- 6 copies of backup data, stored on at least 2 different media types, with 1 copy located offsite
- Ability to recover data backups within 3 hours
- No more than 2 hours of data loss**
Ensuring backups across different media guards against hardware failures, disasters, or human errors contaminating your recovery data copies. The 3 hour maximum recovery window gets businesses back online quickly after outages. And minimizing potential data loss to just 2 hours protects organizations from losing substantial work or transactions during the failure window.
While challenging to fully implement, these guidelines represent best practices for backup redundancy and business continuity.
Now let‘s examine how cloud backups in particular help companies achieve the central tenets of the 6-3-2 rule.
Why Back Up to the Cloud?
Backing up NAS storage offsite to the cloud provides the following advantages:
Air-gapped copies
- Isolates backups from local threats of data corruption, malicious attacks, or onsite disasters
Built-in redundancy
- Cloud platforms replicate your data across multiple physical servers typically across various regions to ensure resilience and availability
Accessibility
- Retrieve files easily from anywhere with an internet connection
Scalability on demand
- Cloud capacity scales massively to accommodate growing backup demands
Cost savings
- Pay for only the storage used rather than maintaining a secondary physical site
According to RightScale‘s 2022 State of the Cloud report, over 90% of enterprises leverage public cloud services with spend increasing annually.
| Year | Average Enterprise Spend on Public Cloud |
|^|^|
| 2019 | $1.97 million |
| 2020 | $2.38 million |
| 2021 | $2.79 million |
| 2022 | $3.10 million |
Driven largely by data protection use cases, backing up critical information in the cloud now serves as a cornerstone to hybrid cloud data management strategies.
Securing Your Data While in the Cloud
While the cloud promises many advantages, it also intrinsically exposes your data to additional risks outside your firewall. Channels of access multiplication on networks you don‘t control directly.
Surveys demonstrate security ranks among the primary challenges companies face when adopting cloud platforms.
Therefore, properly safeguarding the sensitive data you back up in the cloud remains imperative. As a full-stack developer well-versed in cybersecurity, I recommend layered defenses that incorporate:
Encryption
- Encrypt your data first before transferring to cloud servers so no unauthorized party can view underlying content
Access controls
- Leverage role-based access controls, multi-factor authentication, VPN tunnels into cloud accounts, etc.
Key management
- Securely generate encryption keys used to cipher data then store separate from cloud account for added protection
Network segmentation
- Appropriately segment cloud networks away from other environments, which limits attack surface
Intrusion detection
- Monitor account activity for anomalies that could indicate breaches
Backups of backups
- Orchestrate additional backup copies in distinct cloud accounts or physical media
While the cloud furnishes a sturdy shelter for your data during disasters, dedicating resources towards fortifying defenses remains essential not to introduce additional risk.
Next let‘s breakdown key features to configure for robust cloud backups.
Configuring Optimal Cloud Backup Settings
Based on my firsthand experience architecting cloud native applications, I want to share recommendations for tailoring backup settings for best results.
Backup Frequency
- Back up new file changes at least once daily
- Additional intraday backups add minimal overhead if bandwidth allows
Retention Policy
- Maintain weekly backups for one month
- Keep monthly backups for 6-12 months
- Archive annual yearly backups indefinitely
Compression
- Absolutely enable compression to reduce storage volume
- Software compression via LZMA averages ~30% size reduction
- Appliance and source deduplication provide up to 5X higher reduction
Transfer Rate/Bandwidth Utilization
- Lowest acceptable speed: 1 Mbps
- Recommended speed: 15 Mbps+, saturate available bandwidth
- Faster throughput reduces backup window and lowers RPO
Encryption
- 256-bit AES encryption provides robust security
- Encrypt locally before transferring for optimal control
Software vs. Appliance Backups: A Comparison
While the cloud furnishes limitless capacity, upload bandwidth still imposes practical limits. Backing up directly over the Internet works well for smaller data sets. But as storage needs scale, dedicated backup appliances capable of faster throughput prove advantageous.
| Consideration | Software Backup | Appliance Backup |
|---|---|---|
| Cost | Lower | Hardware investment required |
| Speed | Dependent on WAN bandwidth | Load faster with LAN speeds |
| Compression | Limited to software compression | Inline deduplication multiplies reduction |
| Encryption | Software only | Often hardware accelerated |
| Retention | Constrained by cloud capacity | Scale onsite then offload aged snapshots |
Software backups make excellent solutions for smaller workloads. However, onsite appliances both improve performance and offer ample capacity for short term backup snapshots before offloading older data to the cloud on a schedule. This hybrid model reaps optimization advantages from both onsite and cloud resources.
Step-by-Step Guide to Backing Up a Synology NAS
Now that we have covered techniques for safeguarding data and strategic elements for backup schemes, let me provide a walkthrough for backing up a Synology NAS appliance to the cloud:
Prerequisites
- Synology NAS appliance
- Cloud storage account (Google Drive, Dropbox, AWS S3, Azure, Wasabi, etc)
- Hyper Backup package on Synology
Directions
- Install Hyper Backup package
- Launch Hyper Backup and create new backup task
- Select cloud destination and authorize access
- Choose which NAS folders to back up
- Enable compression to reduce size
- Encrypt data before transferring for security
- Schedule daily backup interval
- Configure retention policy to prune backups
- Save settings and run backup task
Initial transfer will take time depending on data volume and upload bandwidth. Subsequent backups will only propagate updated changes.
Make sure to schedule integrity checks of backups for validation. Follow best practices around encryption, isolation, access management to secure your data.
For help troubleshooting or optimizing backup performance, engage an expert managed service provider.
Restoring Backups from the Cloud
To fully validate backups, restoring content remains imperative. When outages strike, promptly recovering data represents the only metric that matters.
Here are tips for smoothly restoring your Synology NAS backups from the cloud:
- Log into Hyper Backup and browse your backup history
- Select the closest recent clean backup version before corruption/loss
- Isolate a test folder and attempt restoring samples first
- Choose a restore destination – original location not recommended
- Initiate recovery workflow
- Verify integrity of restored folders
- If satisfied, restore remainder of content
Potential issues to watch for:
Long restore times
- Retrieve only business critical data first
- Schedule off-peak times for large restores
Metadata synchronization failures
- Try restoring to alternate non-production area
- Update cloud restore tool
Permission mismatch
- Correct Linux ownership/permissions
- Recheck share configurations
Unencrypted data
- Potentially discard compromised backup
- Ensure valid decryption key supplied
Data integrity failures
- Attempt restore from earlier backup
- Rebuild original NAS instead
Seeking help from experienced storage administrators and leverage redundancies makes navigating disasters manageable.
Ongoing Backup Monitoring
Even with tools automating backups, overseeing ongoing health represents a critical responsibility.
Here are best practices I recommend for backup oversight:
Periodic test restores – Quarterly ensure ability to recover from backups, not just copy
generate alerts – For failed backups, retention issues, suspicious access
Inspect backup logs – For errors or performance issues
Validate encryption – Spot check data remains properly ciphered
Monitor capacity – Project future growth against storage
Checksum integrity checks – Detect data corruption issues
Actively monitoring and verifying backups helps avoid scenarios where businesses assume protection but can‘t actually recover when needed most.
Conclusion
Backing up Synology NAS storage to the cloud supplies a versatile means of implementing a cyber resilient data protection scheme. Cloud platforms facilitate highly redundant and instantly accessible offsite copies to guard against catastrophic data disasters.
Carefully configuring backup frequency, retention policies, encryption, and data verification stands necessary to balance recovery objectives with budget requirements. Monitoring health and conducting test restores gives confidence in operational readiness when calamity strikes.
As businesses become increasingly permeated with data, the importance of contingency planning comes into acute focus. Savvy IT administrators make cloud backup management a top priority on the roadmap to insulation from disruptions that will inevitably come. Developing competency with available data protection toolsets leads the way to true organizational resilience.


