On 2017-08-25 08:55, Ferry Toth wrote:
Op Fri, 25 Aug 2017 07:45:44 -0400, schreef Austin S. Hemmelgarn:
On 2017-08-24 17:56, Ferry Toth wrote:
Op Thu, 24 Aug 2017 22:40:54 +0300, schreef Marat Khalili:
We find that typically apt is very slow on a machine with 50 or so
snapshots and raid10. Slow as in probably 10x slower as doing the
same update on a machine with 'single' and no snapshots.
Other operations seem to be the same speed, especially disk
benchmarks do not seem to indicate any performance degradation.
For meaningful discussion it is important to take into account the
fact
Doing daily updates on a desktop is not uncommon and when 3 minutes
become 30 then many would call that meaningful.
I think the more meaningful aspect here is that it's 30 minutes where
persistent storage is liable to be unusable, not necessarily that it's
30 minutes.
Unusable - probably less (depending on your definition)
And depending on both your hardware and how many updates you're handling
(if things are stalling with dpkg calling fsync, it's probably going to
be bad for other stuff too, even if it's still 'usable').
Irritating - yes
Still nothing compared to Windows 10 deciding to download updates right
after you start a round in an FPS or RTS game...
In all seriousness though, I'm kind of used to half-hour daily updates
(or 2 plus hour ones if GCC, LLVM, or glibc are being updated), I run
Gentoo on most of my systems, so everything gets built locally.
Similar for a single office server, which is upgraded twice a year,
where an upgrade normally would take an hour or 2, but now more than a
working day. In the meantime, take samba and postgresql offline,
preventing people to work for a few hours.
That should only be the case if:
1. You don't have your data set properly segregated from the rest of
your system (it should not be part of the upgrade snapshot, but an
independent snapshot taken separately).
2. You are updating the main system, instead of updating the snapshot
you took.
The ideal method of handling an upgrade in this case is:
1. Snapshot the system, but not the data set.
2. Run your updates on the snapshot of the system.
3. Rename the snapshot and the root subvolume so that you boot into the
snapshot.
4. During the next maintenance window (or overnight), shutdown the
system services, snapshot the data set (so you can roll back if the
update screws up the database).
5. Reboot.
That provides minimal downtime, and removes the need to roll-back if the
upgrade fails part way through (you just nuke the snapshot and start
over, instead of having to manually switch to the snapshot and reboot).
Wow, yes that does sound ideal. Is that how you do it?
It's what I've been trying to get working on my Gentoo systems for
months now with varying degrees of success. We do some similar trickery
with updates for our embedded systems where I work (albeit a bit
differently), but we're also directly calling RPM for package installs
instead of using higher level tools like Yum or DNF, so it's easy to
point things somewhere else for the install target. I'm pretty sure
this is essentially what NixOS does for updates too, but I don't think
it has BTRFS snapshot support (they use LVM2), and it's a pain to get
used to the declarative system configuration it uses.
FWIW, it would be even easier if BTRFS supported operation as a lower
layer for OverlayFS, then you could put /etc and similar stuff on
OverlayFS, and the package manager never needs to ask if you want to
update config files because it just puts them on the lower layer.
Now I just need Cananonical to update there installer to take of this.
(that is: tell it to update the system on another partition (subvolume)
than the one mounted on /, and not stop any running system services).
Yeah, that's one of the things I kind of hate about many mainstream
distros. Gentoo is nice in this respect, you can just run the update
command with ROOT=/whatever in the environment, and it will use that for
installing packages, and it expects the admin to handle restarting
services. RPM actually has an option to handle this too these days, but
not Yum or DNF (don't know about Zypper or YaST2, but I think they do
have such an option). dkpg itself probably does, but last I checked,
apt-get, aptitude, and most of the graphical frontends don't.
It's technically possible to work around this using containers (that is,
you spawn a container with the snapshot as it's root filesystem, and run
the update from there), it's just not easy to do right now.
Or run a virtual machine on the server that boots from the snapshot and
that update. Oh no, the virtual machine would slow down my running server
to much.
Depends on how you put it together. This also has issues because of
starting a new instance of init.
Eh, share the snapshot via cifs or nfs to another machine that does a
netboot and let that do the update.
This won't work reliably for other reasons, and will probably kill your
network performance.
O wait, I forgot, I installed btrfs to make our system maintenance easier
than with ext. Maybe that was a mistake, at least until distros take
advantage of the aadvantage and start avoiding the pitfalls?
It really depends on your particular use case.
My point is: fsync is not targeted specifically in many common disk
bench marks (phoronix?), it might be posible that there is no trigger
to spend much time on optimizations in that area. That doesn't make it
meaningless.
that dpkg infamously calls fsync after changing every bit of
information, so basically you're measuring fsync speed. Which is slow
on btrfs (compared to simpler filesystems), but unrelated to normal
work.
OTOH it would be nice if dpkg would at last start making use btrfs
snapshot features and abandon these unnecssary fsyncs completely,
instead restoring a failed install from a snapshot. This would probably
result in a performance improve compared to ext4.
Not dpkg, apt-get and wherever other frontedd you use (although all the
other dpkg frontends I know of are actually apt-get frontends). Take a
look at how SUSE actually does this integration, it's done through
Zypper/YaST2, not RPM. If you do it through dpkg, or RPM, or whatever
other low-level package tool, you need to do a snapshot per package so
that it works reliably, while what you really need is a snapshot per
high-level transaction.
FWIW, if you can guarantee that the system won't crash during an update
(or are actually able to roll back by hand easily if it won't boot), you
can install libeatmydata and LD_PRELOAD it for the apt-get (or aptitude,
or synaptic, or whatever else) call, then call sync afterwards and
probably see a significant perofrmance improvement. The library itself
overloads *sync() calls to be no-ops, so it's not safe to use when you
don't have good fallback options, but it tends to severely improve
performance for stuff like dpkg.
Yeah I can guarantee that it can crash... All you need to do is start the
upgrade from a remote terminal, forget to use screen and then close the
terminal. And if not the installer will run until the first 'do you want
to replace the conf file Y/n' in the background, at which point you have
no choice than to nuke it.
I meant the system itself, not the update. What matters is that the
filesystem isn't shut down unexpectedly, not that the update fails part
way through (if you're using snapshots to give restore points, you've
already got the case of the update failing covered).
But probably if you take a snapshot before eating the data you should be
able to recover.
Indeed.
--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to majord...@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html