On Sunday 23 March 2003 02:50, Steinar H. Gunderson wrote: > But then somebody said each HT `virtual CPU' had > their own part of the bus, so it would definitely help with I/O bound (RAM > I/O, of course, not disk I/O) programs as well... Could this be true, or is > this just misinformation?
I think perhaps someone is getting confused; memory access is chipset-dependent not CPU-dependent. The "Granite Bay" chipsets support interleaved DDR access which doubles the effective bandwidth. On Sunday 23 March 2003 02:13, John R Pierce wrote: > > the newest xeons have 533Mhz bus, which is supported by chipsets like the > E7501. I started running 4 instances of mprime on a pair of dual 2.8Ghz > Xeons, but had to wipe them a few days later and forgot to save the > work-in-progress... Monday I'll restart them and note how fast 1 and 2 > instances run with and without hyperthreading enabled. IIRC, they thought > they'd finish 18,xxx,xxx exponents in 10 days. My 2.66 GHz P4/Asus P4G8X system (e7502 chipset) is running exponent 18600979 (1024K run length) at 0.040 sec/iter giving a total run time of ~8.5 days. Though it uses a "Granite Bay" chipset, this mobo supports "consumer" S478 P4 CPUs. I'm using a "Northwood" 2.66 GHz processor (which doesn't support HT, though the chipset does) because this seems to be optimum grunt/$ at present. > > Note re: the memory contention issue, the dual xeon chipsets like the e7501 > have higher memory bandwidth as they use interleaved DDR (2 banks), this > may at least partially shift the performance vis-a-vis two seperate p4 > systems. Possibly, but dual-bank DDR on a uniprocessor system is better still - puts P4 DDR systems into the same league as systems supporting (expensive) PC1066 RDRAM, maybe even a few percent ahead though using only PC2100 DDR ("266 MHz" actually 133 MHz dual-pumped). > OTOH, dual xeon e7501 systems are not cheap. The ones I built for work > were $3300 each with dual 2.8Ghz and 2GB ram, but without hard drives, > these are 2U rackmount servers using Intel's SE7501WV2 motherboard and a > Intel SE2300 rack chassis. They are also *extremely* noisy (seems to be a > feature of all dual xeon 2U rack servers as they need massive cooling for > the CPUs and 6 hotswap SCSI drives). The availability of consumer mobos with "Granite Bay" chipsets makes Xeon-based systems look _very_ expensive for the CPU power you get from them. Effectively the only performance advantage from the Xeon is the larger L2 cache - memory contention issues will totally undermine this so far as we're concerned. The benefit of Xeon server systems is power density - useful if you want to put a large bundle of them in a small area. But shifting all that heat from a small case really is going to require a lot of airflow, hence the noise. In a 2U rackmount case there's not much height for a heatsink & fan, therefore small components have to be driven fast. Even then the airflow from a rackfull of servers is _warm_ - sufficiently so to be useful as e.g. a hairdryer - you're going to need aircon to dump the excess heat to the outside world. Summary - anyone self-building systems to run Prime95/mprime at home is _almost certainly_ going to get far more CPU power per dollar (purchase price; electricity costs will be similar) from 2 x P4 systems using "Granite Bay" chipset than from 1 x dual Xeon system with the same speed CPU. Regards Brian Beesley _________________________________________________________________________ Unsubscribe & list info -- http://www.ndatech.com/mersenne/signup.htm Mersenne Prime FAQ -- http://www.tasam.com/~lrwiman/FAQ-mers