Re: pxeboot with jumbo frame network
I know I have this working, however I don't remember what I did. I know I can pxeboot and install RHEL on a 9K frame network from a FreeBSD tftp server/NAT gateway. I do know the first thing in my RHEL install script is to set the MTU to 9K. If I have a chance later today, I'll dig into one of my install servers and try to figure out what options I used with DHCP to get it working. Sorry I can't be of more help than to say its possible. Bob Healey Systems Administrator Biocomputation and Bioinformatics Constellation and Molecularium hea...@rpi.edu (518) 276-4407 On 7/23/2012 1:31 PM, Robert Blayzor wrote: Is it possible to PXEboot a diskless client with an MTU higher than 1500 (Ie: 9k) ? The network we currently boot several diskless machines from has a lot of NFS traffic and we'd like to enable them all for larger MTU's however the one thing stopping us is the issue with initial boot seemingly only supporting an MTU of 1500. I know we can set the MTU later (post boot) on the diskless machines, but it doesn't seem we can run a higher MTU on the TFTP server as we'd have TFTP UDP traffic with mis-matched frame sizes at that point. ___ freebsd-stable@freebsd.org mailing list http://lists.freebsd.org/mailman/listinfo/freebsd-stable To unsubscribe, send any mail to freebsd-stable-unsubscr...@freebsd.org
Problems with crashing IBM X3630 M3/ZFS
Hello. I've got a quartet of IBM x3630 M3 with one that is frequently hard locking under heavy NFS load. I am running 9.0-RELEASE with all the patches from freebsd-update. My problem machine has 8 16 core clients, each doing IO intensive tasks connected to it via a Procurve and the onboard igb0 interface. Mostly network reads, typically 10MB read per MB written. When the machine locks under load, none of the consoles respond, nor can I reach the machine via ethernet. I can break into DDB via the serial over lan interface, and am running a debug/witness kernel at the moment (I was running GENERIC previously). During the boot sequence, witness tosses me into DDB ~10 times before I get a login prompt. Prior to this machine acting up, it had multiple 802.1q vlans, and ran 9K packets on its private network to the compute clients. A dmesg can be found at http://boyle.che.rpi.edu/~healer/boomer/dmesg /etc/rc.conf can be found at http://boyle.che.rpi.edu/~healer/boomer/rc.conf A listing of installed ports can be found at http://boyle.che.rpi.edu/~healer/boomer/pkg_info The output of psauxwwo wchan against my two crash dumps can be found at http://boyle.che.rpi.edu/~healer/boomer/crash1-psaux-wchan and http://boyle.che.rpi.edu/~healer/boomer/crash2-psaux-wchan I'm not entire convinced this is software, but I've run out of local experts to ask, and can't prove its hardware. -- Bob Healey Systems Administrator Biocomputation and Bioinformatics Constellation and Molecularium hea...@rpi.edu (518) 276-4407 ___ freebsd-stable@freebsd.org mailing list http://lists.freebsd.org/mailman/listinfo/freebsd-stable To unsubscribe, send any mail to freebsd-stable-unsubscr...@freebsd.org
Re: about thumper aka sun fire x4500
The X4500s are oldish systems built around a pair Opteron 290 chips with 16 GB RAM and 6 PCI-X Marvell SATA controllers with 8 ports each supporting 48 drives in the machine. Only the first and 4th drive on the I think 4th controller are bootable. Are you using the latest firmware? If not, you're going to have to pay Oracle for the privilege of updating it as there is no way the machines are still under warranty. I'd find a copy of the OpenSolaris Live CD and see if that boots and supports all your drives. Hope this helps you, or helps someone else on the list with more knowledge of debugging older AMD systems point you in the right direction. Bob Healey Systems Administrator Biocomputation and Bioinformatics Constellation and Molecularium hea...@rpi.edu (518) 276-4407 On 1/17/2012 5:09 PM, Jeremy Chadwick wrote: On Tue, Jan 17, 2012 at 06:59:08PM +0100, peter h wrote: I have been beating on of these a few days, i have udes freebsd 9.0 and 8.2 Both fails when i engage 10 disks, the system craches and messages : Hyper transport sync flood will get into the BIOS errorlog ( but nothing will come to syslog since reboot is immediate) Using a zfs radz of 25 disks and typing zpool scrub will bring the system down in seconds. Anyone using a x4500 that can comfirm that it works ? Or is this box broken ? I do not have one of these boxes / am not familiar with them, but HyperTransport is an AMD thing. The concept is that it's a bus that interconnects different pieces of a system to the CPU (and thus the memory bus). ASCII diagram coming up: +---+ | RAM | +--++ | +--++ | CPU (w/ on-die MCH) | +--++ | +--++ +-+ | HyperTransport bridge +-+ PCI Express bus (VGA, etc.) | +--++ +-+ | +--+---+ | Southbridge (SATA, etc.) | +--+ ZFS is memory I/O intensive. Your controller, given that it consists of 25 disks, is probably sitting on the PCI Express bus, and thus is generating an equally high amount of I/O. Given this above diagram, I'm sure you can figure out how flooding might occur. :-) I'm not sure what sync flood means (vs. I/O flooding). Googling turns up *tons* of examples of this on the web, except every time they involve people doing overclocking or having CPU-level problems pertaining to voltage. There may be a BIOS option on your system to help curb this behaviour, or at least try to limit it in some way. I know on our AMD systems at work the number of options in the Memory section of the BIOS is quite large, many of which pertaining to interactivity with HyperTransport. If you want my advice? Bring the issue up to Sun. They will almost certainly be able to assign the case to an engineer, who although may not be familiar with FreeBSD, hopefully WILL be familiar with the bus interconnects described above and might be able to help you out. ___ freebsd-stable@freebsd.org mailing list http://lists.freebsd.org/mailman/listinfo/freebsd-stable To unsubscribe, send any mail to freebsd-stable-unsubscr...@freebsd.org