I searched for dif and apparently the Linux kernel has supported this type of data integrity checks.
Interesting read about it in the docs.
Just be careful; although they say its part of mainline there could be gotchas. No one uses sata drives in enterprise setting. So these features would be well tested for sas drives and such. I remember we had so many regressions on our storage array when t10dif was enabled. But that was long time ago. I just realized that this was mainline after clicking the link you provided.
Does raid really protect data? I've always heard raid is not backup.
Funny this guy completely fumbled. Got confused between checksum and parity. And you can definitely recover raid when controller dies, there is a common practice of labeling the drives with stickers for this exact reason. His explanation of raid 5 is totally wrong. (raid 5 stores parity not checksum)
This could be good guide to raid interview questions… but…
0 - stripe
1 - mirror
5 - parity
10 - mirror while striping
50 - mirror, stripe, parity
60 i think is double parity. There are more distributed raid types.
Now coming to spinning disks to wards the end of the video… we are currently working on something like raid on enterprise nvme drives. So raid only for spinning disks is bollocks.
Also, there is one more technology. Its called erasure coding. Very popular in the cloud because it requires 1.4 storage to achieve something like raid 10. So money… if you want ec at home; use ceph-fs on a raspberry pi for home nas systems.
Also, modern systems not only raid but also create storage tiers. Like using nvme for cache etc…you can try all of that on ceph. Just create a kubernetes instance of ceph and ceph dashboard will let you configure all of that.