Hi, We tried turning on switch control on our local cluster (www.dcsc.sdu.dk) but were unable to achieve any improvement in scale up whatsoever. I was wondering if you folks could shed light upon how we should go ahead with this. (We have not installed the all-to-all patch yet)
The cluster architecture is as follows: ########## * Computing nodes 160x Dell PowerEdge 1950 1U rackmountable servers with 2 2,66Ghz Intel Woodcrest CPUs, 4 GB Ram, 2x160 GB HDD (7200rpm, 8 MB buffer, SATA150), 2x Gigabit Ethernet 40x Dell PowerEdge 1950 1U rackmountable servers with 2 2,66Ghz Intel Woodcrest CPUs, 8 GB Ram, 2x160 GB HDD (7200rpm, 8 MB buffer, SATA150), 2x Gigabit Ethernet ########## * Switches 9 D-link SR3324 2 D-link SRi3324 The switches are organised in two stacks, each connected to the infrastracture switch with an 8 Gb/s LACP trunk.Firmware Build on the switches ########## * Firmware Build on the switches: 3.00-B16 There are newer firmware builds available, but according to the update logs, there is not update on the IEEE flow control protocol in the new firmware ########## * Tests (were run using OPENMPI, not LAMMPI) DPPC-bilayer system of ~ 40000 atoms, with PME and cutoffs, 1fs time step. The scaleup data is as follows. We are also currently running some tests with larger systems. # Procs nanoseconds/day Scaleup 1 0.526 1 2 1.0 1.90 4 1.768 3.36 8 1.089 2.07 16 0.39 0.74 Any inputs will be very helpful, thank you Best, -himanshu _______________________________________________ gmx-users mailing list gmx-users@gromacs.org http://www.gromacs.org/mailman/listinfo/gmx-users Please search the archive at http://www.gromacs.org/search before posting! Please don't post (un)subscribe requests to the list. Use the www interface or send it to [EMAIL PROTECTED] Can't post? Read http://www.gromacs.org/mailing_lists/users.php