On 2017-10-18 09:53, Peter Grandi wrote:
I forget sometimes that people insist on storing large
volumes of data on unreliable storage...
Here obviously "unreliable" is used on the sense of storage that
can work incorrectly, not in the sense of storage that can fail.
Um, in what world is a device randomly dropping off the bus (this is the
primary issue with USB for storage) not a failure? Yes, it's not a
catastrophic failure (for BTRFS at least), and it's transient (the
kernel will re-enumerate the device when it resets the bus), but that
doesn't change the fact that the service that is supposed to be provided
by the device failed.
To clarify more concretely, when I say 'unreliable' in reference to
computers technology (and for that matter, almost anything else), I mean
something that encounters non-trivial error states, either correctable
or uncorrectable, at a frequency above that which is deemed reasonable
for the designed function of the device.
In my opinion the unreliability of the storage is the exact
reason for wanting to use raid1. And I think any problem one
encounters with an unreliable disk can likely happen with more
reliable ones as well, only less frequently, so if I don't
feel comfortable using raid1 on an unreliable medium then I
wouldn't trust it on a more reliable one either.
Oh please, please a bit less silliness would be welcome here.
In a previous comment on this tedious thread I had written:
> If the block device abstraction layer and lower layers work
> correctly, Btrfs does not have problems of that sort when
> adding new devices; conversely if the block device layer and
> lower layers do not work correctly, no mainline Linux
> filesystem I know can cope with that.
> Note: "work correctly" does not mean "work error-free".
The last line is very important and I added it advisedly.
You seem to be using "unreliable" in two completely different
meanings, without realizing it, as both "working incorrectly"
and "reporting a failure". They are really very different.
And you seem to be using the term 'failure' to only mean 'catastrophic
failure'. Strictly speaking, even that is 'working incorrectly', albeit
in a much more specific and permanent manner than just returning errors.
Even looking at things that way though, Zoltan's assessment that
reliability is essentially a measure of error rate is correct. Internal
SATA devices absolutely can randomly drop off the bus just like many USB
storage devices do, but it almost never happens (it's a statistical
impossibility if there are no hardware or firmware issues), so they are
more reliable in that respect.
The "working incorrectly" general case is the so called
"bizantine generals problem" and (depending on assumptions) it
is insoluble.
Btrfs has some limited ability to detect (and sometimes recover
from) "working incorrectly" storage layers, but don't expect too
much from that.
--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to majord...@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html