Hello Ralph, > Could you please clarify - are you mixing 32 and 64 bit versions > in your runs that have a problem?
No, I have four different versions on each machine. tyr fd1026 1250 ls -ld /usr/local/openmpi-1.6.5_* drwxr-xr-x 7 root root 512 May 23 14:00 /usr/local/openmpi-1.6.5_32_cc drwxr-xr-x 7 root root 512 May 23 13:55 /usr/local/openmpi-1.6.5_32_gcc drwxr-xr-x 7 root root 512 May 23 10:12 /usr/local/openmpi-1.6.5_64_cc drwxr-xr-x 7 root root 512 May 23 12:21 /usr/local/openmpi-1.6.5_64_gcc "/usr/local" is a link to machine specific files on a NFS server. lrwxrwxrwx 1 root root 25 Jan 10 07:47 local -> /export2/prog/SunOS_sparc lrwxrwxrwx 1 root root 26 Oct 5 2012 local -> /export2/prog/SunOS_x86_64 ... I can choose a package in my file "$HOME/.cshrc". tyr fd1026 1251 more .cshrc ... #set MPI = openmpi-1.6.5_32_cc #set MPI = openmpi-1.6.5_32_gcc #set MPI = openmpi-1.6.5_64_cc #set MPI = openmpi-1.6.5_64_gcc ... source /opt/global/cshrc ... "/opt/global/cshrc" determines the processor architecture and operating system and calls package specific initialization files. tyr fd1026 1258 more /opt/global/mpi.csh ... case openmpi-1.6.5_32_cc: case openmpi-1.6.5_32_gcc: case openmpi-1.6.5_64_cc: case openmpi-1.6.5_64_gcc: ... if (($MPI == openmpi-1.7_32_cc) || ($MPI == openmpi-1.9_32_cc) || \ ($MPI == ompi-java_32_cc) || ($MPI == ompi-java_32_gcc) || \ ($MPI == openmpi-1.7_32_gcc) || ($MPI == openmpi-1.9_32_gcc)) then if ($JDK != jdk1.7.0_07-32) then echo " " echo "In '${MPI}' funktioniert 'mpijavac' nur mit" echo "'jdk1.7.0_07-32'. Waehlen Sie bitte das entsprechende" echo "Paket in der Datei '${HOME}/.cshrc' aus und melden Sie" echo "sich ab und wieder an, wenn Sie 'mpiJava' benutzen" echo "wollen." echo " " endif endif ... setenv OPENMPI_HOME ${DIRPREFIX_PROG}/$MPI ... set path = ( $path ${OPENMPI_HOME}/bin ) ... Sorry for the german message in my shell script, but mpi.csh sets all necessary environment variables for the selected package. I must logout and login again, if I select a different package in "$HOME/.cshrc", so that I never mix environments for different packages, because my home directory and "/opt/global" are the same on all machines (they are provided via an NFS server). > If that isn't the case, then the error message is telling you that > the system thinks you are mixing optimized and debug versions - > i.e., one node is using an optimized version of OMPI and another > is using a debug version. This also isn't allowed. I build my packages with copy-paste from a file. All configure commands use "--enable-debug" (three different architectures with two different compilers each). tyr openmpi-1.6.5 1263 grep -- enable-debug README-OpenMPI-1.6.5 --enable-debug \ --enable-debug \ --enable-debug \ --enable-debug \ --enable-debug \ --enable-debug \ tyr openmpi-1.6.5 1264 > If you check and find those two conditions are okay, then I suspect > you are hitting the Solaris "bit rot" problem that we've talked > about before - and are unlikely to be able to fix any time soon. sunpc1 hello_1 113 mpiexec -mca btl ^udapl -np 4 -host sunpc1 hello_1_mpi Process 2 of 4 running on sunpc1 ... sunpc1 hello_1 114 mpiexec -mca btl ^udapl -np 4 -host linpc1 hello_1_mpi [sunpc1:05035] [[4165,0],0] ORTE_ERROR_LOG: Buffer type (described vs non-described) mismatch - operation not allowed in file ../../../../../openmpi-1.6.5a1r28554/orte/mca/grpcomm/bad/grpcomm_bad_module.c at line 841 ^Cmpiexec: killing job... I get the following table, if I use every machine as local machine and run my command on one of my hosts. local machine | -host | | sunpc1 | linpc1 | rs1 -----------------------------+--------+--------+------- sunpc1 (Solaris 10, x86_64) | ok | hangs | hangs linpc1 (Solaris 10, x86_64) | hangs | ok | ok rs1 (Solaris 10, sparc) | hangs | ok | ok It seems that I have a problem with Solaris x86_64 and gcc-4.8.0, if I use a 64-bit version of Open MPI. I have no problems with Sun C and a 64-bit version of Open MPI or any 32-bit version of Open MPI. Do you have any idea, what I can do to track the problem and to get a solution? Kind regards Siegmar > On May 24, 2013, at 12:02 AM, Siegmar Gross <siegmar.gr...@informatik.hs-fulda.de> wrote: > > > Hi > > > > I installed openmpi-1.6.5a1r28554 on "openSuSE Linux 12.1", "Solaris 10 > > x86_64", and "Solaris 10 sparc" with gcc-4.8.0 and "Sun C 5.12" in 32- > > and 64-bit versions. Unfortunately I have a problem with the 64-bit > > version, if I build Open MPI with gcc. The program hangs and I have > > to terminate it with <Ctrl-c>. > > > > > > sunpc1 hello_1 144 mpiexec -mca btl ^udapl -np 4 \ > > -host sunpc1,linpc1,rs0 hello_1_mpi > > [sunpc1:15576] [[16182,0],0] ORTE_ERROR_LOG: Buffer type (described vs > > non-described) mismatch - operation not allowed in file > > ../../../../../openmpi-1.6.5a1r28554/orte/mca/grpcomm/bad/grpcomm_bad_module.c > > at line 841 > > ^Cmpiexec: killing job... > > > > sunpc1 hello_1 145 which mpiexec > > /usr/local/openmpi-1.6.5_64_gcc/bin/mpiexec > > sunpc1 hello_1 146 > > > > > > I have no problems with the 64-bit version, if I compile Open MPI > > with Sun C. Both 32-bit versions (compiled with "cc" or "gcc") work > > as expectedas well. > > > > sunpc1 hello_1 106 mpiexec -mca btl ^udapl -np 4 \ > > -host sunpc1,linpc1,rs0 hello_1_mpi > > Process 2 of 4 running on rs0.informatik.hs-fulda.de > > Process 0 of 4 running on sunpc1 > > Process 3 of 4 running on sunpc1 > > Process 1 of 4 running on linpc1 > > Now 3 slave tasks are sending greetings. > > Greetings from task 3: > > message type: 3 > > msg length: 116 characters > > message: > > hostname: sunpc1 > > operating system: SunOS > > release: 5.10 > > processor: i86pc > > ... > > > > sunpc1 hello_1 107 which mpiexec > > /usr/local/openmpi-1.6.5_64_cc/bin/mpiexec > > > > > > > > sunpc1 hello_1 106 mpiexec -mca btl ^udapl -np 4 \ > > -host sunpc1,linpc1,rs0 hello_1_mpi > > Process 2 of 4 running on rs0.informatik.hs-fulda.de > > Process 3 of 4 running on sunpc1 > > Process 0 of 4 running on sunpc1 > > Process 1 of 4 running on linpc1 > > ... > > > > sunpc1 hello_1 107 which mpiexec > > /usr/local/openmpi-1.6.5_32_gcc/bin/mpiexec > > > > > > I would be grateful, if somebody can fix the problem for the > > 64-bit version with gcc. Thank you very much for any help in > > advance. > > > > > > Kind regards > > > > Siegmar > > > > _______________________________________________ > > users mailing list > > us...@open-mpi.org > > http://www.open-mpi.org/mailman/listinfo.cgi/users > >