> I don't know all states of this file system, and copies you have. Right now 
> the earliest copy is obviously broken, and the latest copy is probably more 
> broken because at the least its csum tree has been blown away meaning there's 
> no checksums to confirm whether any data extracted/copied from the file 
> system is OK. You'd just have to open the file and look at it to see if it 
> behaves as it should, and even then depending on what kind of file it is, 
> corruption may not be obvious immediately. So... catch22.

I'm not sure I follow you. I have a clean dd image, and I made a fresh
copy of it every time I tried something new, so no copy should be any
more broken than whatever the current recovery operation did to it.
(There shouldn't be a distinction between earliest and latest if I
understand what you mean by "copy".)

I get all those tree errors when it tries to access the "active"
versions of the @ and @home volumes. But by the time it starts
restoring the snapshots I don't see those anymore. So my
interpretation of the output was that the trees rooted at @ and @home
are messed up, but the tree structures for the snapshots are ok. Am I
misunderstanding that output, or are we both equally in the dark about
what it's telling us?

Setting aside the tree integrity issue, does anyone know what
condition causes that "offset is <number>" output during a restore
operation? I've seen several wiki pages and posts about how to use the
restore feature, but nothing that details what's going on under the
hood or how to interpret the output. If it's not complaining about the
tree, and there are none of those "offset" errors, would the
assumption be that the entire snapshot was recovered successfully?

>But seems to me in any case sda6 needs to be reformatted, the only question is 
>if you reformat it Btrfs or something else (honestly - it doesn't matter 
>whether it was the chicken soup that made you sick, the association is 
>understandable).

Yep. I'll probably stick with Btrfs and just try to improve my backup
strategy. I think I need to start a) replicating my snapshots to
another file system and b) more aggressively snapshotting @home. I've
been avoiding that because I have so many big files that come and go
so often there the snapshots would take up a lot of space. But maybe a
single rolling snapshot hourly or daily would make this kind of issue
easier to deal with.

> I'm very interesting in how this turns out. Is it a Btrfs bug, is it a 
> hardware bug, is it some bad/unlucky collision of more than one problem? Is 
> it possible metadata DUP might have changed the outcome? If so are we better 
> off in the short term using metadata DUP on SSDs? Or is the hindsight takeway 
> "metadata DUP isn't worth it on SSD, better is more frequent backups than 
> what was done" and you just have to weigh whether that's practical for your 
> workflow.

I'll definitely report back here if #btrfs provides any insight. And
anyone else who comes across this and has suggestions, please, chime
in!
--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to