On Fri, 13 Jan 2023 11:17:29 -0500
John Blinka <john.bli...@gmail.com> wrote:

> On Thu, Jan 12, 2023 at 12:54 PM Laurence Perkins <lperk...@openeye.net>
> wrote:
> 
> > I’m not sure if I’m doing something horribly wrong, or missing something
> > blindingly obvious, but I’ve just had to boot a rescue shell yet again, so
> > I’m going to ask.
> >
> >
> >
> > To save time and effort, I have my big, powerful machine create
> > binpackages for everything when it updates, and then let all my smaller
> > machines pull from that.  It works pretty well for the most part.
> >  
> 
> I do something quite similar, but have never had a glibc problem. Maybe the
> problem lies in differences between the specific details of our two
> approaches.
> 
> I have 3 boxes with different hardware but identical portage setup,
> identical world file, identical o.s., etc, even identical CFLAGS, CPPFLAGS
> and CPU_FLAGS_X86 despite different processors. Like you, I build on my
> fastest box (but offload work via distcc), and save binpkgs. After a world
> update (emerge -DuNv —changed-deps @world) , I rsync all repositories and
> binpkgs from the fast box to the others. An emerge -DuNv —changed-deps
> —usepkgonly @world on the other boxes completes the update. I do this
> anywhere from daily to (rarely) weekly. Portage determines when to update
> glibc relative to other packages. There hasn’t been a problem in years with
> glibc.
> 
> I believe there are more sophisticated ways to supply updated portage trees
> and binary packages across a local network.  I think there are others on
> the list using these more sophisticated techniques successfully. Just a
> plain rsync satisfies my needs.
> 
> It’s not clear to me whether you have the problem on  your big powerful
> machine or on your other machines. If it’s the other machines, that
> suggests that portage knows the proper build sequence on the big machine
> and somehow doesn’t on the lesser machines. Why? What’s different?
> 
> Perhaps there’s something in my update frequency or maintaining an
> identical setup on all my machines that avoids the problem you’re having?
> 
> If installing glibc first works, then maybe put a wrapper around your
> emerge? Something that installs glibc first if there’s a new binpkg then
> goes on to the remaining updates.
> 
> Just offered in case there’s a useful hint from my experience - not arguing
> that mine is the one true way (tm).
> 
> HTH,
> 
> John Blinka
> 
> >  

In case it is not clear what the underlying problem is:

Slow machine updates and is on the same set of packages as the fast
machine.

Fast machine updates glibc to a new version at time T1
Fast machine updates app-arch/xz-utils to a new version at time T2.
This version of xz CAN have glibc symbols from the very newest glibc
version that was merged at time T1. Everything is fine on the fast
machine.

Now the slow machine starts its update process at a time T3 > T2. The
list of packages includes glibc AND xz-utils, however xz-utils is often
pulled in before glibc which ends in a disaster.
Now you have an xz decompressing tool on your slow machine that cannot
run, because some library symbols from glibc are missing (because
glibc was not merged yet), and you're pretty much stuck in the middle
of the update with a broken system.

I have seen this kind of behaviour only when I have not updated for a
very long time the slow machine (i.e. no update for a year).

Anyway I think a reasonable default for emerge would be to merge glibc
as early as possible, because all other binary packages could have
been built with the newer glibc version, and could potentially
fail to run on the slow machine until glibc is updated.

Hope that clears up what happens, and why it fails to update / breaks
the slow machine.

Andreas

Reply via email to