Thanks for the advice, we tried "-mca btl_openib_ib_min_rnr_timer 25
-mca btl_openib_ib_timeout 20" but we are still getting errors as we
increase the Ns of HPL.dat value into the thousands. Is it ok to just
add these valuse to .openmpi/mca-params.conf for the user running the
test or should we add these setting to each node in
/usr/local/etc/openmpi-mca-params.conf
The OpenFabrics stack has reported a network error event. Open MPI
will try to continue, but your job may end up failing.
Local host: compute-0-8.local
MPI process PID: 30544
Error number: 10 (IBV_EVENT_PORT_ERR)
This error may indicate connectivity problems within the fabric;
please contact your system administrator.
Ralph Castain wrote:
The default retry values are wrong and will be corrected in the next
OMPI release. For now, try running with:
-mca btl_openib_ib_min_rnr_timer 25 -mca btl_openib_ib_timeout 20
Should work.
Ralph
On Mar 26, 2009, at 2:16 PM, Gary Draving wrote:
Hi Everyone,
I'm doing some performance testing using HPL with TCP turned off. My
HPL.dat file looks like the following:
It seems to work well for lower Ns values but as I increase that
value it inevitably fails with
"[[13535,1],169][btl_openib_component.c:2905:handle_wc] from
compute-0-0.local to: compute-0-8 error polling LP CQ with status
RETRY EXCEEDED ERROR status number 12 for wr_id 209907960 opcode 0
qp_idx 3"
The machines in this test are all dual core quads with built in
mellanox cards for total of 320 processors
It seems like once I reach a certain nuber of "Ns" the infiniban
starts having problems. Increasing "btl_openib_ib_retry_count" and
"btl_openib_ib_timeout" to the max allowed us to get from 4096 to
8192 Ns but the error came back at around 8192.
If anyone has some ideas on this problem I would be very interests,
Thanks
((((((((((((((((((HPL.dat file being uses )))))))))))))))))))
HPLinpack benchmark input file
Innovative Computing Laboratory, University of Tennessee
HPL.out output file name (if any)
6 device out (6=stdout,7=stderr,file)
1 # of problems sizes (N)
8192 Ns
1 # of NBs
256 NBs
0 PMAP process mapping (0=Row-,1=Column-major)
1 # of process grids (P x Q)
19 Ps
19 Qs
(defaults for rest)
(((((((((((((((((( Full error printout ))))))))))))))))))
[[13535,1],169][btl_openib_component.c:2905:handle_wc] from
compute-0-0.local to: compute-0-8 error polling LP CQ with status
RETRY EXCEEDED ERROR status number 12 for wr_id 209907960 opcode 0
qp_idx 3
--------------------------------------------------------------------------
The InfiniBand retry count between two MPI processes has been
exceeded. "Retry count" is defined in the InfiniBand spec 1.2
(section 12.7.38):
The total number of times that the sender wishes the receiver to
retry timeout, packet sequence, etc. errors before posting a
completion error.
This error typically means that there is something awry within the
InfiniBand fabric itself. You should note the hosts on which this
error has occurred; it has been observed that rebooting or removing a
particular host from the job can sometimes resolve this issue.
Two MCA parameters can be used to control Open MPI's behavior with
respect to the retry count:
* btl_openib_ib_retry_count - The number of times the sender will
attempt to retry (defaulted to 7, the maximum value).
* btl_openib_ib_timeout - The local ACK timeout parameter (defaulted
to 10). The actual timeout value used is calculated as:
4.096 microseconds * (2^btl_openib_ib_timeout)
See the InfiniBand spec 1.2 (section 12.7.34) for more details.
Below is some information about the host that raised the error and the
peer to which it was connected:
Local host: compute-0-0.local
Local device: mthca0
Peer host: compute-0-8
You may need to consult with your system administrator to get this
problem fixed.
--------------------------------------------------------------------------
--------------------------------------------------------------------------
mpirun has exited due to process rank 169 with PID 26725 on
node compute-0-0 exiting without calling "finalize". This may
have caused other processes in the application to be
terminated by signals sent by mpirun (as reported here).
--------------------------------------------------------------------------
_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users
_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users