On redhat 7.1 this command still works:

[ec2-user@ip-10-0-0-175 ~]$ cat /etc/redhat-release
Red Hat Enterprise Linux Server release 7.1 (Maipo)

[ec2-user@ip-10-0-0-175 ~]$ grep processor /proc/cpuinfo | wc -l
4


On Mon, Mar 7, 2016 at 9:56 AM, Steve Varnau <[email protected]> wrote:

> > It seems that the parallel make fails on 8 GB machines.
>
> I think your first sentence overstates the determinism of the problem a
> bit.
> I ran a normal, default build on 8GB machine last week and had no problem.
> There must be an environmental problem, but I don't think we fully
> understand it yet.
>
> The aggressiveness of the make parallelism is set in core/sqf/sqenvcom.sh.
> It sets the parallel factor based on how many CPUs are on your machine:
>
> # Set default build parallelism
> # Can be overridden on make commandline
> cpucnt=$(grep processor /proc/cpuinfo | wc -l)
> #     no number means unlimited, and will swamp the system
> export MAKEFLAGS="-j$cpucnt"
>
> If that calculation is wrong, maybe that could cause a problem.
>
> --Steve
>
>
> > -----Original Message-----
> > From: Gunnar Tapper [mailto:[email protected]]
> > Sent: Monday, March 7, 2016 9:35 AM
> > To: [email protected]
> > Subject: Parallel Make Failures
> >
> > Hi,
> >
> > It seems that the parallel make fails on 8 GB machines. At least, Nitin
> > and
> > I both ran into make failures that did not appear when running serial
> > make.
> > I've also seen similar failures when building the code on 12 GB machines.
> >
> > Based on previous discussions, the Trafodion Contributor Guide recommends
> > rerunning make a few times if running issues.
> >
> > I most wonder if there's a way to reduce the aggressiveness of the make
> in
> > general. Could we, for example, come up with a table that correlates
> > system
> > size to define the -l option or something similar?
> >
> > --
> > Thanks,
> >
> > Gunnar
> > *If you think you can you can, if you think you can't you're right.*
>



-- 
Thanks,

Amanda Moran

Reply via email to