On 11/13/22 13:02, hw wrote:
On Fri, 2022-11-11 at 07:55 -0500, Dan Ritter wrote:hw wrote:On Thu, 2022-11-10 at 20:32 -0500, Dan Ritter wrote:Linux-Fan wrote: [...] * RAID 5 and 6 restoration incurs additional stress on the other disks in the RAID which makes it more likely that one of them will fail. The advantage of RAID 6 is that it can then recover from that...Disks are always being stressed when used, and they're being stessed as well when other types of RAID arrays than 5 or 6 are being rebuild. And is there evidence that disks fail *because* RAID arrays are being rebuild or would they have failed anyway when stressed?Does it matter? The observed fact is that some notable proportion of RAID 5/6 rebuilds fail because another drive in that group has failed.Fortunately, I haven't observed that. And why would only RAID 5 or 6 be affected and not RAID 1 or other levels?
Any RAID level can suffer additional disk failures while recovering from a disk failure. I saw this exact scenario on my SOHO server in August 2022. The machine has a stripe of two mirrors of two HDD's each (e.g. ZFS equivalent of RAID10). One disk was dying, so I replaced it. While the replacement disk was resilvering, a disk in the other mirror started dying. I let the first resilver finish, then replaced the second disk. Thankfully, no more disks failed. I got lucky.
David