Re: Confirm that bnx2i iSCSI offload works for RHEL 5.5 on DELL EqualLogic
On 08/24/2010 10:13 AM, Michael Wolf wrote: Hi ALL, I try to set up oracle cluster with four nodes (DELL R610) and EqualLogig PS6000. Servers has four Broadcom 5709C dual network cards, two intenal, two PCI cards. I`m trying to get the bnx2i driver to work using iSCSI offload for four channels. I can't get it to work, getting error iscsiadm: Could not login to [iface: bnx2i.00:22:19:68:3a:04, target: iqn.2001-05.com.equallogic:0-8a0906-76c630e04-e0492d35ddc63f-a1test, portal: 10.10.77.10,3260): iscsiadm: initiator reported error (4 - encountered connection failure) Showing nothing in /var/log/messages and no events on the EQ box. So there is no iferror error in messages or nothing from iscsid? DELL suport team is not shure if bnx2i driver will work with RHEL 5.5, they ask me to reinstall RHEL 5.2, from my opinion is that not an option. I am using RHEL 5.5, oracle linux, kernel 2.6.18-194.11.1.0.1.el5 on Is that a oracle linux kernel or the rhel 5.5 one? Are you running rhel 5.5 in a oracle linux vm? x86_64 iscsi-initiator-utils-6.2.0.871-0.16.el5 Also you need the updated iscsi utils. http://people.redhat.com/mchristi/iscsi/rhel5.6/iscsi-initiator-utils/ All firmware of the R610 is up to date All interfaces are shown as bnx2i [r...@ora-db-rac1-04 ~]# scsiadm -m iface -bash: scsiadm: command not found [r...@ora-db-rac1-04 ~]# iscsiadm -m iface default tcp iser iser bnx2i.00:22:19:68:3a:04 bnx2i,00:22:19:68:3a: 04,10.10.77.69,, Are these IPs on the same subnet as the network port they are attached to? bnx2i.00:22:19:68:3a:0a bnx2i,00:22:19:68:3a:0a, 10.10.77.68,, bnx2i.00:00:00:00:00:00 bnx2i, 00:00:00:00:00:00,,, bnx2i.00:22:19:68:3a:06 bnx2i,00:22:19:68:3a: 06,10.10.77.67,, bnx2i.00:10:18:64:2e:5b bnx2i,00:10:18:64:2e:5b, 10.10.77.66,, bnx2i.00:10:18:64:98:bb bnx2i,00:10:18:64:98:bb, 10.10.77.65,, bnx2i.00:22:19:68:3a:08 bnx2i,00:22:19:68:3a: 08,10.10.77.64,, bnx2i.00:10:18:64:98:b9 bnx2i, 00:10:18:64:98:b9,10.10.77.63,, Only four of them are connected to the SAN. How to figure out which hw- adress belongs to which interface? Start the iscsi service (or at least make sure bnx2i is loaded) then run: [r...@meanminna kernel]# iscsiadm -m host -P 1 Host Number: 8 State: running Transport: bnx2i Initiatorname: IPaddress: HWaddress: 00:10:18:15:fd:f1 Netdev: eth1 You can see the network device the iscsi offload interface is attached to (in this case eth1), and you can see the MAC/hwaddress for the iscsi interface. Is bnx2i driver supported for RHEL5.5? Yes. Is there a more detailed description for setting up such an environment as the RedHat manuals? Section 5.1.2 Setting up a iface for a iSCSI offload card of the iscsi readme in /usr/share/docs/iscsi-initiator-utils-$VERSION/README -- You received this message because you are subscribed to the Google Groups "open-iscsi" group. To post to this group, send email to open-is...@googlegroups.com. To unsubscribe from this group, send email to open-iscsi+unsubscr...@googlegroups.com. For more options, visit this group at http://groups.google.com/group/open-iscsi?hl=en.
Re: iscsi seems to work, but no device node?
On 08/24/2010 11:43 AM, agshekeloh wrote: Hi, I'm running open-iscsi 2.0-871, on Ubuntu 10.04 amd64, straight from the operating system package. Mounting an iSCSI drive works perfectly when booting from a hard drive, but not when booting diskless via NFS. The iSCSI server is OpenSolaris. The diskless node can discover the iSCSI node and log in: # iscsiadm -m session -P1 Target: iqn.1986-03.com.sun:02:2293f889-f1bb-e764-e45f-b7931c6c86a5 Current Portal: XXX.XXX.64.168:3260,1 Persistent Portal: XXX.XXX.64.168:3260,1 ** Interface: ** Iface Name: default Iface Transport: tcp Iface Initiatorname: iqn.1993-08.org.debian: 01:cd00198dd976 Iface IPaddress: XXX.XXX.199.20 Iface HWaddress: Iface Netdev: SID: 1 iSCSI Connection State: LOGGED IN iSCSI Session State: LOGGED_IN Internal iscsid Session State: NO CHANGE I would expect to see /dev/sda at this point, but it doesn't appear. What does iscsiadm -m session -P 3 print out? Could you also send the /var/log/messages output when you run the login command that fails to find disks. Also if you do echo - - - > /sys/class/scsi_host/hostX/scan X is the host number you will see from the -P3 command. Does that find disks (check /var/log/messages and /dev after the echo has completed). -- You received this message because you are subscribed to the Google Groups "open-iscsi" group. To post to this group, send email to open-is...@googlegroups.com. To unsubscribe from this group, send email to open-iscsi+unsubscr...@googlegroups.com. For more options, visit this group at http://groups.google.com/group/open-iscsi?hl=en.
iscsi seems to work, but no device node?
Hi, I'm running open-iscsi 2.0-871, on Ubuntu 10.04 amd64, straight from the operating system package. Mounting an iSCSI drive works perfectly when booting from a hard drive, but not when booting diskless via NFS. The iSCSI server is OpenSolaris. The diskless node can discover the iSCSI node and log in: # iscsiadm -m session -P1 Target: iqn.1986-03.com.sun:02:2293f889-f1bb-e764-e45f-b7931c6c86a5 Current Portal: XXX.XXX.64.168:3260,1 Persistent Portal: XXX.XXX.64.168:3260,1 ** Interface: ** Iface Name: default Iface Transport: tcp Iface Initiatorname: iqn.1993-08.org.debian: 01:cd00198dd976 Iface IPaddress: XXX.XXX.199.20 Iface HWaddress: Iface Netdev: SID: 1 iSCSI Connection State: LOGGED IN iSCSI Session State: LOGGED_IN Internal iscsid Session State: NO CHANGE I would expect to see /dev/sda at this point, but it doesn't appear. There's no nodes starting with "ip" in /dev/disk/by-path, either. I ran iscsid -fd8 by hand and looked for errors, but it looked much like the output from the working machine (at least to my untrained eye). Any suggestions on how to debug this further? Thanks, ==ml -- You received this message because you are subscribed to the Google Groups "open-iscsi" group. To post to this group, send email to open-is...@googlegroups.com. To unsubscribe from this group, send email to open-iscsi+unsubscr...@googlegroups.com. For more options, visit this group at http://groups.google.com/group/open-iscsi?hl=en.
Confirm that bnx2i iSCSI offload works for RHEL 5.5 on DELL EqualLogic
Hi ALL, I try to set up oracle cluster with four nodes (DELL R610) and EqualLogig PS6000. Servers has four Broadcom 5709C dual network cards, two intenal, two PCI cards. I`m trying to get the bnx2i driver to work using iSCSI offload for four channels. I can't get it to work, getting error iscsiadm: Could not login to [iface: bnx2i.00:22:19:68:3a:04, target: iqn.2001-05.com.equallogic:0-8a0906-76c630e04-e0492d35ddc63f-a1test, portal: 10.10.77.10,3260): iscsiadm: initiator reported error (4 - encountered connection failure) Showing nothing in /var/log/messages and no events on the EQ box. DELL suport team is not shure if bnx2i driver will work with RHEL 5.5, they ask me to reinstall RHEL 5.2, from my opinion is that not an option. I am using RHEL 5.5, oracle linux, kernel 2.6.18-194.11.1.0.1.el5 on x86_64 iscsi-initiator-utils-6.2.0.871-0.16.el5 All firmware of the R610 is up to date All interfaces are shown as bnx2i [r...@ora-db-rac1-04 ~]# scsiadm -m iface -bash: scsiadm: command not found [r...@ora-db-rac1-04 ~]# iscsiadm -m iface default tcp iser iser bnx2i.00:22:19:68:3a:04 bnx2i,00:22:19:68:3a: 04,10.10.77.69,, bnx2i.00:22:19:68:3a:0a bnx2i,00:22:19:68:3a:0a, 10.10.77.68,, bnx2i.00:00:00:00:00:00 bnx2i, 00:00:00:00:00:00,,, bnx2i.00:22:19:68:3a:06 bnx2i,00:22:19:68:3a: 06,10.10.77.67,, bnx2i.00:10:18:64:2e:5b bnx2i,00:10:18:64:2e:5b, 10.10.77.66,, bnx2i.00:10:18:64:98:bb bnx2i,00:10:18:64:98:bb, 10.10.77.65,, bnx2i.00:22:19:68:3a:08 bnx2i,00:22:19:68:3a: 08,10.10.77.64,, bnx2i.00:10:18:64:98:b9 bnx2i, 00:10:18:64:98:b9,10.10.77.63,, Only four of them are connected to the SAN. How to figure out which hw- adress belongs to which interface? Is bnx2i driver supported for RHEL5.5? Is there a more detailed description for setting up such an environment as the RedHat manuals? I am searching for the best way to figure out he reason for the connect failures? Thanks for any helping hand. -- You received this message because you are subscribed to the Google Groups "open-iscsi" group. To post to this group, send email to open-is...@googlegroups.com. To unsubscribe from this group, send email to open-iscsi+unsubscr...@googlegroups.com. For more options, visit this group at http://groups.google.com/group/open-iscsi?hl=en.
Re: detected conn error (1011)
Hi, On Fri, 2010-08-06 at 15:57 +0100, Hannes Reinecke wrote: > Mike Christie wrote: > > ccing Hannes from suse, because this looks like a SLES only bug. > > > > Hey Hannes, > > > > The user is using Linux 2.6.27 x86 based on SLES + Xen 3.4 (as dom0) > > running a couple of RHEL 5.5 VMs. The underlying storage for these VMs > > is iSCSI based via open-iscsi 2.0.870-26.6.1 and a DELL equallogic array. > > > > > > On 08/05/2010 02:21 PM, Goncalo Gomes wrote: > >> I've copied both the messages file from the host goncalog140 and the > >> patched libiscsi.c. FWIW, I've also included the iscsid.conf. Find these > >> files in the link below: > >> > >> http://promisc.org/iscsi/ > >> > > > > It looks like this chunk from libiscsi.c:iscsi_queuecommand: > > > > case ISCSI_STATE_FAILED: > > reason = FAILURE_SESSION_FAILED; > > sc->result = DID_TRANSPORT_DISRUPTED << 16; > > break; > > > > is causing IO errors. > > > > You want to use something like DID_IMM_RETRY because it can be a long > > time between the time the kernel marks the state as ISCSI_STATE_FAILED > > until we start recovery and properly get all the device queues blocked, > > so we can exhaust all the retries if we use DID_TRANSPORT_DISRUPTED. > Yeah, I noticed. > But the problem is that multipathing will stall during this time, > ie no failover will occur and I/O will stall. Using DID_TRANSPORT_DISRUPTED > will circumvent this and we can failover immediately. > > Sadly I got additional bugreports about this so I think I'll have > to revert it. I applied and tested the changes Mike Christie suggests. After the LUN is rebalanced within the array I no longer see the IO errors and it appears the setup is now resilient to the equallogic LUN failover process. I'm attaching the log from the dmesg merely for sanity check purposes, if anyone cares to take a look? > I have put some test kernels at > > http://beta.suse.com/private/hare/sles11/iscsi Do the test kernels in the url above contain the change of DID_TRANSPORT_DISRUPTED to DID_DIMM_RETRY or is there more to it than simply changing the result code? If the latter, would you be able to upload the source rpms or a unified patch containing the changes you are are staging? I'm looking for a more pallatable way to test them, given I have no SLES box lying around, but will install one if needs be. Thanks, -Goncalo. -- You received this message because you are subscribed to the Google Groups "open-iscsi" group. To post to this group, send email to open-is...@googlegroups.com. To unsubscribe from this group, send email to open-iscsi+unsubscr...@googlegroups.com. For more options, visit this group at http://groups.google.com/group/open-iscsi?hl=en. device-mapper: multipath: version 1.0.5 loaded device-mapper: multipath round-robin: version 1.0.0 loaded device-mapper: table: 251:1: multipath: error getting device device-mapper: ioctl: error adding target to table device-mapper: table: 251:1: multipath: error getting device device-mapper: ioctl: error adding target to table Citrix Systems, Inc. -- Private Release Kernel Private File Disclaimer The private files provided to you contain a preliminary code fix. These private files have been created and distributed to you to address your specific issue and provide Citrix with the feedback that your issue has been resolved or to provide further debugging information. These private files have had minimal in-house testing with no regression testing and may contain defects. These private file(s) will only be supported until an official Hotfix has been provided or one is publicly available from the Citrix web site. Any private files that are provided to you are intended only for the use of the individual or entity to which this is addressed and distribution of these files or utilities is prohibited. CITRIX MAKES NO REPRESENTATIONS OR WARRANTIES OF NONINFRINGEMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE WITH RESPECT TO THE PRIVATE FILES. THE PRIVATE FILES ARE DELIVERED ON AN "AS IS" BASIS. YOU SHALL HAVE THE SOLE RESPONSIBILITY FOR ADEQUATE PROTECTION AND BACK-UP OF AN<6>Loading iSCSI transport class v2.0-870. iscsi: registered transport (tcp) scsi6 : iSCSI Initiator over TCP/IP connection1:0: detected conn error (1011) scsi 6:0:0:0: Direct-Access EQLOGIC 100E-00 4.3 PQ: 0 ANSI: 5 sd 6:0:0:0: [sdb] 209725440 512-byte hardware sectors: (107 GB/100 GiB) sd 6:0:0:0: [sdb] Write Protect is off sd 6:0:0:0: [sdb] Mode Sense: ad 00 00 00 sd 6:0:0:0: [sdb] Write cache: disabled, read cache: enabled, doesn't support DPO or FUA sd 6:0:0:0: [sdb] 209725440 512-byte hardware sectors: (107 GB/100 GiB) sd 6:0:0:0: [sdb] Write Protect is off sd 6:0:0:0: [sdb] Mode Sense: ad 00 00 00 sd 6:0:0:0: [sdb] Write cache: disabled, read cache: enabled, doesn't support DPO or FUA sdb: sdb1 sd 6:0:0:0: [sdb] Attached SCSI disk sd 6:0:0:0: Attached scsi generic sg1 type 0 tap_backend_changed: