[Dovecot] Better to use a single large storage server or multiple smaller for mdbox?
Ed W
lists at wildgooses.com
Thu Apr 12 13:58:52 EEST 2012
On 12/04/2012 11:20, Stan Hoeppner wrote:
> On 4/11/2012 9:23 PM, Emmanuel Noobadmin wrote:
>> On 4/12/12, Stan Hoeppner<stan at hardwarefreak.com> wrote:
>>> On 4/11/2012 11:50 AM, Ed W wrote:
>>>> One of the snags of md RAID1 vs RAID6 is the lack of checksumming in the
>>>> event of bad blocks. (I'm not sure what actually happens when md
>>>> scrubbing finds a bad sector with raid1..?). For low performance
>>>> requirements I have become paranoid and been using RAID6 vs RAID10,
>>>> filesystems with sector checksums seem attractive...
>>> Except we're using hardware RAID1 here and mdraid linear. Thus the
>>> controller takes care of sector integrity. RAID6 yields nothing over
>>> RAID10, except lower performance, and more usable space if more than 4
>>> drives are used.
>> How would the control ensure sector integrity unless it is writing
>> additional checksum information to disk? I thought only a few
>> filesystems like ZFS does the sector checksum to detect if any data
>> corruption occurred. I suppose the controller could throw an error if
>> the two drives returned data that didn't agree with each other but it
>> wouldn't know which is the accurate copy but that wouldn't protect the
>> integrity of the data, at least not directly without additional human
>> intervention I would think.
> When a drive starts throwing uncorrectable read errors, the controller
> faults the drive and tells you to replace it. Good hardware RAID
> controllers are notorious for their penchant to kick drives that would
> continue to work just fine in mdraid or as a single drive for many more
> years. The mindset here is that anyone would rather spent $150-$2500
> dollars on a replacement drive than take a chance with his/her valuable
> data.
>
I'm asking a subtlely different question.
The claim by ZFS/BTRFS authors and others is that data silently "bit
rots" on it's own. The claim is therefore that you can have a raid1 pair
where neither drive reports a hardware failure, but each gives you
different data? I can't personally claim to have observed this, so it
remains someone else's theory... (for background my experience is
simply: RAID10 for high performance arrays and RAID6 for all my personal
data - I intend to investigate your linear raid idea in the future though)
I do agree that if one drive reports a read error, then it's quite easy
to guess which pair of the array is wrong...
Just as an aside, I don't have a lot of failure experience. However,
the few I have had (perhaps 6-8 events now) is that there is a massive
correlation in failure time with RAID1, eg one pair I had lasted perhaps
2 years and then both failed within 6 hours of each other. I also had a
bad experience with RAID 5 that wasn't being scrubbed regularly and when
one drive started reporting errors (ie lack of monitoring meant it had
been bad for a while), the rest of the array turned out to be a
patchwork of read errors - linux raid then turns out to be quite fragile
in the presence of a small number of read failures and it's extremely
difficult to salvage the 99% of the array which is ok due to the disks
getting kicked out... (of course regular scrubs would have prevented
getting so deep into that situation - it was a small cheap nas box
without such features)
Ed W
More information about the dovecot
mailing list