On Sun, Mar 02, 2014 at 05:47:36PM -0600, Bruce Dubbs wrote:
> Ken, You mentioned a problem with linux-3.13.5.  Can you explain your 
> issues with that version?
> 
>    -- Bruce

 Ah!  I've just put a short summary in my reply to you re libexecdir
on blfs-dev.  Here's a MUCH fuller one:

 I built 7.5 with a 3.13.5 kernel (and 3.13.5 headers) to test some
gnome packages.  I hadn't built those packages since
October/November and there was quite a lot of change amongst them,
so I kept the new system in chroot (with a 3.13.4 kernel) until I'd
finished libreoffice.  Up to then, everything was fine.

 I then booted it - still ok - and started building the remainder.
I'd forgotten that my original "build everything else" script (from
when I was testing that everything built with make-4.0) included
java/icedtea, and because I'd forgotten that, those versions were out
of date.  Realised it was building them, thought "oh well, let it
build the old versions, no benefit but no harm done" and left it
running.  When I woke up in the morning, I was surprised to find
that my OpenJDK script was still running rm -rf for the source
directory, and had been doing so for more than 5 hours (both
wall-clock time and CPU time), and was now at 99%-100% of one CPU,
according to top.

 This build was running as root (yeah, I know) and somehow
/etc/passwd- seemed to have been altered during OpenJDK.
Specifically, I couldn't su or login, and none of the previous
package's logs showed that file had changed during the package's
build.  But I also need to mention that this is an AMD phonon
running at -j4 for the builds, and it tends to "lose its lunch"
doing that.  In fact, pass-1 gcc usually does this (an ICE), and
dropping the caches is not usually sufficiebnt to allow it to
proceed.  So, much of LFS was built with -j3.  I think I had changed
back to -j4 for docbook and xorg.

 Anyway, I booted an older system, chrooted, fixed up the passwords.
After that I resumed my build - got as far as GConf, and again rm
-rf was taking 100% of one cpu, but this time only for about 15
minutes before I killed it.  Built 3.13.4, rebooted, everything else
built fine running 3.13.4.

 Tried SysRQ-T before killing the rm -rf, but rm was not running,
merely ready to run.  So, I asked on lkml about trying to find out
what is going on.  I got a suggestion to use perf (thought I hadn't
enabled that, but in fact I had done or perhaps it is a default on
x86), and to perhaps use 'crash' which needs a kernel with debugging
information, and a saved vmlinux.  So, I turned on a *lot* of
debugging options, built crash, booted it - the kernel was noticeably
slower, even simple things showed they were using a lot of CPU in
top, but I couldn't replicate the problem.

 Yesterday (Sunday) evening, I went back to the regular 3.13.5
kernel, confirmed that perf did work there, and tried to replicate
the problem without any success.  Then I built the new system - if
nothing else, I'll be able to update and test my kde scripts on this
build.  Last time I looked, it was still running ok at -j4.

 Once again, I'm starting to think that cosmic rays are the problem,
or perhaps someone forgot to feed or muck-out the imps, or someone
upset the camel. [ Can you guess what I'm reading at the moment ? ].

 Probably, this kernel is fine for most people.  Even if I can
replicate the problem, it might be something in my .config.  OTOH,
this does mean it's going to be some days before I get round to
testing 3.14-rc kernels :-(  As always, YMMV and you may be burned
even by a stable kernel.

ĸen
-- 
das eine Mal als Tragödie, dieses Mal als Farce
-- 
http://linuxfromscratch.org/mailman/listinfo/lfs-dev
FAQ: http://www.linuxfromscratch.org/faq/
Unsubscribe: See the above information page

Reply via email to