A lot of people hear that RAID "protects your data" and stop there. That sounds reassuring, but it's only partly true.
RAID can help keep a system online when a drive fails, and in some setups it can also improve storage performance. What it can't do is protect you from every kind of data loss. If files are deleted, corrupted, encrypted by ransomware, or overwritten by mistake, RAID usually won't save you. That's why it's important to understand what RAID is actually for, what each RAID level does, and how it compares to things like ZFS, software RAID, hardware RAID, and backups.
In this guide, we'll break down RAID in plain English, compare the main RAID levels, explain when each one makes sense, and cover where ZFS fits into the conversation.
What RAID is
RAID stands for Redundant Array of Independent Disks. The idea is simple: instead of treating each drive as a separate device, RAID combines multiple drives into one logical storage unit.
Depending on the RAID level, the array may spread data across drives for speed, duplicate data for redundancy, or store parity information so the array can survive one or more failed disks.
That means RAID is usually built around one or both of these goals:
Performance, by splitting reads and writes across multiple drives
Fault tolerance, by keeping enough duplicate or parity information to recover from a failed drive
Different RAID levels make different tradeoffs. Some focus almost entirely on speed. Others focus more on redundancy. Some do a bit of both, but with more complexity.
How RAID protects your data
RAID protects data by reducing the chance that a single drive failure takes down the entire system.
For example, if you use a mirrored setup and one disk dies, the other disk still has the same data. If you use a parity-based RAID setup, the array can rebuild the missing data from the remaining drives and parity information.
This matters for servers, virtualization hosts, databases, and storage systems where uptime matters. A failed drive doesn't have to mean immediate downtime or emergency recovery from scratch.
Still, this protection has limits.
RAID does not replace backups. It doesn't protect against accidental deletion, file corruption, malware, controller failure in some setups, fire, theft, or a bad command run by a human. If bad data gets written to the array, RAID usually protects that bad write just as faithfully as it would a good one.
So the better way to think about RAID is this: RAID helps with drive failure and availability. Backups help with actual disaster recovery.
RAID vs backups
This is the most important distinction in the whole article.
RAID is about keeping storage available when hardware fails. Backups are about restoring data after something goes wrong.
If a drive dies in a RAID 1 mirror, the server can keep running. Great. But if someone deletes a client folder and that deletion syncs instantly across the array, RAID has done exactly what it was supposed to do, and your files are still gone.
A real data protection strategy usually includes both:
RAID for uptime and drive failure tolerance
Backups for recovery from deletion, corruption, ransomware, and larger disasters
That combination is much safer than relying on either one alone.
How RAID works
RAID works through a few core techniques: striping, mirroring, and parity.
Striping means splitting data across multiple disks. This can improve performance because multiple drives can read or write parts of the same workload at the same time.
Mirroring means storing identical copies of data on more than one disk. This gives you redundancy, but reduces usable capacity because the same data is written multiple times.
Parity means storing calculated data that can be used to reconstruct missing information if a disk fails. This is more space-efficient than full mirroring, but it adds complexity and can slow down writes.
Every RAID level is basically a different combination of these ideas.
Comparing all RAID levels
Not every RAID level is common in modern deployments, but it's still useful to understand the full landscape.
RAID 0
RAID 0 uses striping only. Data is split across two or more drives, which can improve read and write performance.
The catch is that RAID 0 has no redundancy at all. If one drive fails, the entire array is lost because parts of the data were stored across all disks.
RAID 0 gives you the full combined capacity of all drives. Two 2 TB drives in RAID 0 give you 4 TB usable. That's attractive for speed and space, but it's risky.
RAID 0 is best for temporary data, scratch space, or workloads where performance matters more than fault tolerance. It is not a good choice for important data.
RAID 1
RAID 1 uses mirroring. Every write is copied to at least two drives.
This means if one drive fails, the other drive still contains the full dataset. RAID 1 is simple, reliable, and easy to understand. It doesn't give you the same capacity efficiency as parity-based RAID, but it is a common choice for operating system volumes, boot drives, and smaller servers.
The tradeoff is usable space. Two 2 TB drives in RAID 1 give you only 2 TB usable, because one drive is effectively the mirror of the other.
RAID 1 is a good fit when simplicity and redundancy matter more than raw capacity.
RAID 2
RAID 2 is largely historical and almost never used in modern systems. It used bit-level striping and special error correction methods across multiple disks.
You usually won't see RAID 2 offered in real-world server deployments today.
RAID 3
RAID 3 uses byte-level striping with a dedicated parity disk. It can tolerate a single drive failure, but the dedicated parity disk becomes a bottleneck, especially for writes.
Like RAID 2, RAID 3 is mostly of historical interest now.
RAID 4
RAID 4 uses block-level striping with a dedicated parity disk. It improved on some of the limitations of RAID 3, but it still suffers from that dedicated parity bottleneck.
Modern systems generally prefer RAID 5 or RAID 6 instead.
RAID 5
RAID 5 uses block-level striping with distributed parity. Instead of storing parity on one dedicated disk, parity data is spread across all drives in the array.
This avoids the single parity-disk bottleneck found in RAID 4 and allows the array to survive one disk failure.
RAID 5 needs at least three drives. Usable capacity is roughly the total capacity minus one drive's worth. For example, four 2 TB drives in RAID 5 give you about 6 TB usable.
RAID 5 became very popular because it offers a decent balance of performance, redundancy, and space efficiency. But it has real downsides on larger arrays and bigger disks. Rebuilds can take a long time, and during that period the array is vulnerable. If a second drive fails before rebuild finishes, the array is gone.
For modern large-capacity disks, many admins are much more cautious about RAID 5 than they used to be.
RAID 6
RAID 6 is similar to RAID 5, but it stores double distributed parity instead of single parity. That means the array can survive two drive failures instead of one.
It requires at least four drives, and usable capacity is roughly total capacity minus two drives' worth. Four 2 TB drives in RAID 6 give you about 4 TB usable.
RAID 6 is often preferred over RAID 5 for larger arrays because the added protection is worth the extra capacity cost. The tradeoff is slower writes and less usable storage.
If you're working with big SATA drives and want parity-based redundancy, RAID 6 is often the safer choice.
RAID 10
RAID 10, sometimes written as RAID 1+0, combines mirroring and striping. It stripes data across mirrored pairs of drives.
This gives you strong read and write performance, along with better redundancy than RAID 0 and usually faster rebuild behavior than parity-based RAID. It requires at least four drives.
With four 2 TB drives in RAID 10, you get about 4 TB usable. Like RAID 1, half of the raw capacity is used for redundancy.
RAID 10 is widely used for virtualization hosts, databases, and production systems where performance and resilience both matter. The main downside is capacity efficiency. You give up more usable space than with RAID 5 or RAID 6.
RAID 50
RAID 50 combines multiple RAID 5 groups with striping across them. It can improve performance and fault tolerance compared to a single RAID 5 set, but it's more complex and usually found in larger enterprise storage setups.
It still inherits RAID 5's parity and rebuild concerns, just in a more layered design.
RAID 60
RAID 60 combines multiple RAID 6 groups with striping across them. Like RAID 50, it is generally used in larger storage environments where you want more scale and more fault tolerance.
This can survive multiple drive failures depending on where the failures occur, but complexity and overhead both increase.
Quick RAID level comparison
| RAID level | Minimum drives | Fault tolerance | Performance | Usable capacity |
|---|
| RAID 0 | 2 | None | High | 100% |
| RAID 1 | 2 | 1 drive per mirror | Good reads, decent writes | 50% |
| RAID 2 | Varies | Historical | Historical | Rarely used |
| RAID 3 | 3 | 1 drive | Good sequential, limited by parity disk | Total minus 1 disk |
| RAID 4 | 3 | 1 drive | Better than RAID 3 in some cases, still parity bottleneck | Total minus 1 disk |
| RAID 5 | 3 | 1 drive | Good reads, slower writes | Total minus 1 disk |
| RAID 6 | 4 | 2 drives | Good reads, slower writes than RAID 5 | Total minus 2 disks |
| RAID 10 | 4 | Depends on which drives fail | Very good | 50% |
| RAID 50 | 6+ | Varies by subgroup | High | Better than RAID 10, less than RAID 0 |
| RAID 60 | 8+ | Higher than RAID 50 | High | Lower than RAID 50 |
Software RAID vs hardware RAID
One of the biggest practical questions is whether to use software RAID or hardware RAID.
What software RAID is
Software RAID is managed by the operating system. On Linux, this is often done with tools like mdadm or with filesystems and volume managers that include RAID-like features.
The main advantage is flexibility. Software RAID is often easier to inspect, migrate, automate, and recover, especially in Linux environments. It also avoids dependence on a specific RAID controller model. Modern CPUs are usually more than capable of handling the overhead for many workloads.
Software RAID is also often cheaper because you don't need a dedicated RAID card.
What hardware RAID is
Hardware RAID uses a dedicated controller card or onboard RAID chipset to manage the array independently of the OS.
This used to be much more appealing when CPU overhead mattered more and dedicated RAID cards brought features like battery-backed cache. In some enterprise environments, hardware RAID is still used for specific workflows and vendor-supported storage stacks.
But hardware RAID has downsides too. If the RAID controller fails, recovery may depend on finding a compatible replacement. Management can also be more opaque compared to software RAID, and lower-end "fake RAID" implementations can create more confusion than value.
Which is better?
For many modern Linux servers, software RAID is the more practical choice. It's transparent, portable, and usually easier to troubleshoot.
Hardware RAID can still make sense in some enterprise setups, especially where a specific platform is already standardized around it, or where controller-based caching and vendor support are part of the design.
In other words, this isn't about one option always being better. It's about choosing the simpler and more maintainable tool for the environment you're running.
Where ZFS fits in
ZFS is not traditional RAID, but it often comes up in the same conversation because it can do many of the things people want from RAID, while also adding filesystem-level features.
OpenZFS combines storage management and the filesystem in a way that lets it handle redundancy, snapshots, checksums, and data integrity together. Instead of thinking only about RAID levels, ZFS users often think in terms of vdevs and storage pools.
Common ZFS layouts include mirrors, RAIDZ1, RAIDZ2, and RAIDZ3.
RAIDZ1 is roughly comparable to RAID 5 in that it can tolerate one disk failure
RAIDZ2 is roughly comparable to RAID 6 and can tolerate two disk failures
RAIDZ3 can tolerate three disk failures
Where ZFS stands out is data integrity. ZFS checksums data and metadata, and it can detect silent corruption, often called bit rot. In redundant configurations, it can often repair corrupted data automatically by reading a good copy elsewhere in the pool.
That's a major reason people choose ZFS for storage servers, backup systems, and environments where data correctness matters as much as uptime.
Still, ZFS isn't automatically the right answer for every system. It has its own memory expectations, operational habits, and design choices. It also doesn't remove the need for backups. Snapshots are helpful, but snapshots are not the same thing as off-system backups.
RAID vs ZFS
RAID and ZFS are often discussed as if they are direct competitors, but that oversimplifies things.
Traditional RAID is usually focused on block device redundancy and performance. ZFS is a filesystem and volume manager with integrated redundancy options.
If you want a simple mirrored boot volume or an mdadm RAID 10 for a Linux host, traditional software RAID may be the straightforward option.
If you want end-to-end checksumming, snapshots, pooled storage, and self-healing behavior in redundant configurations, ZFS may be the better fit.
A lot depends on the workload, the operating system, your team's familiarity with the platform, and how much complexity you want to manage.
How to choose the right RAID level
The right RAID level depends on what you're optimizing for.
If you care most about speed and don't care about redundancy, RAID 0 is the obvious answer, though it should be used very carefully.
If you want simple redundancy for a smaller setup or boot volume, RAID 1 is often enough.
If you want a good balance of capacity and protection for larger storage, RAID 6 is often safer than RAID 5.
If you want strong performance and fault tolerance for production workloads, RAID 10 is a common favorite.
If you're building a storage server and care about data integrity features beyond classic RAID, ZFS deserves a serious look.
The biggest mistake is choosing based on raw usable capacity alone. Saving extra space isn't very helpful if rebuild times, failure risk, or recovery complexity become a problem later.
How to set up RAID the right way
The exact steps depend on your OS, controller, and storage stack, but the general process looks similar across most environments.
First, decide what problem you're solving. Are you trying to keep a hypervisor online after a drive failure? Improve database IOPS? Build a storage server? The answer affects whether RAID 1, RAID 6, RAID 10, or ZFS makes the most sense.
Next, use matching drives whenever possible. Mixing capacities and performance levels usually leads to wasted space or uneven behavior.
Then, decide whether you want software RAID, hardware RAID, or ZFS. For many Linux servers, software RAID is the cleaner option. For storage-focused setups, ZFS may be worth the extra planning.
After that, monitor drive health. RAID is not something you set once and forget forever. You want SMART monitoring, alerts, and a plan for replacing failed disks quickly.
And finally, back everything up anyway. This is the part people skip until they regret it.
Common RAID myths
One of the biggest myths is that RAID is a backup. It isn't.
Another common myth is that RAID 5 is always enough. It used to be the default recommendation for a lot of arrays, but today's larger disks and longer rebuild times make that more questionable in many scenarios.
It's also a mistake to assume hardware RAID is always better than software RAID. On modern Linux systems, software RAID is often easier to manage and recover.
And while ZFS has a strong reputation for data integrity, it still isn't magic. You still need sound backups, monitoring, and operational discipline.
Conclusion
RAID is one of those storage topics that sounds more complicated than it really is once you break it down. At its core, it's just a way to combine drives for better performance, redundancy, or both.
The important thing is knowing what kind of protection you're actually getting. RAID can help a server stay online when a disk fails, but it doesn't replace backups, and it doesn't protect against every kind of data loss. For many workloads, RAID 1, RAID 6, and RAID 10 are the most practical traditional choices, while ZFS is worth considering when data integrity and filesystem-level features matter just as much as redundancy.
If you're planning infrastructure for storage-heavy applications, virtualization, backups, or production hosting, the right setup depends on your workload, your recovery goals, and how much complexity you want to manage. xTom provides dedicated servers, colocation, IP transit, shared hosting, and general IT services for a wide range of projects, while V.PS offers scalable NVMe-powered KVM VPS hosting for lighter deployments and flexible cloud workloads.
Ready to discuss your infrastructure needs? Contact the team to explore the right solution for your projects.
Frequently asked questions about RAID
What is RAID in simple terms?
RAID is a way to combine multiple drives into one storage setup to improve performance, add redundancy, or both. Different RAID levels handle that in different ways.
Does RAID protect against data loss?
RAID helps protect against drive failure, but not all data loss. It usually won't protect against accidental deletion, ransomware, corruption, or site-level disasters. That's why you still need backups.
Which RAID level is best?
There isn't one best RAID level for everything. RAID 1 is simple and reliable, RAID 6 is often better for larger parity-based arrays, and RAID 10 is a strong choice when you want both performance and redundancy.
Is RAID 5 still worth using?
RAID 5 can still make sense in some smaller arrays, but many admins are more cautious about it now because rebuild times on large disks can be long, and the array only tolerates one drive failure.
Is RAID 10 better than RAID 5?
For many production workloads, RAID 10 is often preferred because it offers better write performance and usually less stressful rebuild behavior. RAID 5 is more space-efficient, though, so it depends on whether performance or usable capacity matters more.
Is software RAID better than hardware RAID?
Not always, but software RAID is often the simpler and more flexible option on modern Linux servers. Hardware RAID can still make sense in some enterprise environments, especially where a specific controller platform is already part of the design.
Is ZFS better than RAID?
ZFS isn't just a RAID alternative, it's a filesystem and volume manager with integrated redundancy options. It can be a better fit when you want snapshots, checksumming, pooled storage, and protection against silent corruption, but it still requires planning and it still doesn't replace backups.
Can RAID replace backups?
No. RAID helps with uptime after disk failure. Backups help you recover data after deletion, corruption, malware, or larger disasters. You usually want both.