Re: [Lustre-discuss] Getting weird disk errors, no apparent impact

2010-08-13 Thread LaoTsao 老曹
https://cds.sun.com/is-bin/INTERSHOP.enfinity/WFS/CDS-CDS_SMI-Site/en_US/-/USD/viewproductdetail-start?productref=rdacdvr60002500-09.03.0bc02.0013-lx-...@cds-cds_smi rdac On 8/13/2010 12:05 PM, LaoTsao wrote: U mean stk 2540? Iirc one can download drivers from oracle sun site --- Original

Re: [Lustre-discuss] Getting weird disk errors, no apparent impact

2010-08-13 Thread Wojciech Turek
Hi, I don't think you should use rdac path checker in your multipath.conf. I would suggest to use tur pathchecker path_checkertur Bes gerads, Wojciech On 13 August 2010 16:51, David Noriega wrote: > We have three Sun StorageTek 2150, one connected to the metadata > server and two

Re: [Lustre-discuss] Getting weird disk errors, no apparent impact

2010-08-13 Thread LaoTsao
U mean stk 2540? Iirc one can download drivers from oracle sun site --- Original message --- > From: David Noriega > To: lustre-discuss@lists.lustre.org > Sent: 13.8.'10, 11:51 > > We have three Sun StorageTek 2150, one connected to the metadata > server and two crossconnected to the tw

Re: [Lustre-discuss] Getting weird disk errors, no apparent impact

2010-08-13 Thread David Noriega
We have three Sun StorageTek 2150, one connected to the metadata server and two crossconnected to the two data storage nodes. They are connected via fiber using the qla2xxx driver that comes with CentOS 5.5. The multipath daemon has the following config: defaults { udev_dir

Re: [Lustre-discuss] Getting weird disk errors, no apparent impact

2010-08-13 Thread Wojciech Turek
Hi David, I have seen simmilar errors given out by some storage arrays. There were caused by arrays exporting volumes via more then a single path without multi path driver installed or configured properly. Some times the array controllers requires a special driver to be installed on Linux host (fo

[Lustre-discuss] Getting weird disk errors, no apparent impact

2010-08-12 Thread David Noriega
We just setup a lustre system, and all looks good, but there is this nagging error thats floating about. When I reboot any of the nodes, be it a OSS or MDS, I will get this: [r...@meta1 ~]# dmesg | grep sdc sdc : very big device. try to use READ CAPACITY(16). SCSI device sdc: 4878622720 512-byte h