Re: Confirm that bnx2i iSCSI offload works for RHEL 5.5 on DELL EqualLogic

2010-08-24 Thread Mike Christie

On 08/24/2010 10:13 AM, Michael Wolf wrote:

Hi ALL,

I try to set up oracle cluster with four nodes (DELL R610) and
EqualLogig PS6000.
Servers has four Broadcom 5709C dual network cards, two intenal, two
PCI cards.
I`m trying to get the bnx2i driver to work using iSCSI offload for
four channels.
I can't get it to work, getting error
iscsiadm: Could not login to [iface: bnx2i.00:22:19:68:3a:04, target:
iqn.2001-05.com.equallogic:0-8a0906-76c630e04-e0492d35ddc63f-a1test,
portal: 10.10.77.10,3260):
iscsiadm: initiator reported error (4 - encountered connection
failure)

Showing nothing in /var/log/messages and no events on the EQ box.


So there is no iferror error in messages or nothing from iscsid?



DELL suport team is not shure if bnx2i driver will work with RHEL 5.5,
they ask me to reinstall RHEL 5.2, from my opinion is that not an
option.
I am using RHEL 5.5, oracle linux, kernel 2.6.18-194.11.1.0.1.el5 on


Is that a oracle linux kernel or the rhel 5.5 one? Are you running rhel 
5.5 in a oracle linux vm?




x86_64
iscsi-initiator-utils-6.2.0.871-0.16.el5


Also you need the updated iscsi utils.
http://people.redhat.com/mchristi/iscsi/rhel5.6/iscsi-initiator-utils/


All firmware of the R610 is up to date

All interfaces are shown as bnx2i
[r...@ora-db-rac1-04 ~]# scsiadm -m iface
-bash: scsiadm: command not found
[r...@ora-db-rac1-04 ~]# iscsiadm -m iface
default tcp
iser iser
bnx2i.00:22:19:68:3a:04 bnx2i,00:22:19:68:3a:
04,10.10.77.69,,


Are these IPs on the same subnet as the network port they are attached to?


bnx2i.00:22:19:68:3a:0a bnx2i,00:22:19:68:3a:0a,
10.10.77.68,,
bnx2i.00:00:00:00:00:00 bnx2i,
00:00:00:00:00:00,,,
bnx2i.00:22:19:68:3a:06 bnx2i,00:22:19:68:3a:
06,10.10.77.67,,
bnx2i.00:10:18:64:2e:5b bnx2i,00:10:18:64:2e:5b,
10.10.77.66,,
bnx2i.00:10:18:64:98:bb bnx2i,00:10:18:64:98:bb,
10.10.77.65,,
bnx2i.00:22:19:68:3a:08 bnx2i,00:22:19:68:3a:
08,10.10.77.64,,
bnx2i.00:10:18:64:98:b9 bnx2i,
00:10:18:64:98:b9,10.10.77.63,,

Only four of them are connected to the SAN. How to figure out which hw-
adress belongs to which interface?


Start the iscsi service (or at least make sure bnx2i is loaded) then run:

[r...@meanminna kernel]# iscsiadm -m host -P 1
Host Number: 8
State: running
Transport: bnx2i
Initiatorname: 
IPaddress: 
HWaddress: 00:10:18:15:fd:f1
Netdev: eth1

You can see the network device the iscsi offload interface is attached 
to (in this case eth1), and you can see the MAC/hwaddress for the iscsi 
interface.




Is bnx2i driver supported for RHEL5.5?


Yes.


Is there a more detailed description for setting up such an
environment as the RedHat manuals?


Section 5.1.2 Setting up a iface for a iSCSI offload card of the iscsi 
readme in /usr/share/docs/iscsi-initiator-utils-$VERSION/README


--
You received this message because you are subscribed to the Google Groups 
"open-iscsi" group.
To post to this group, send email to open-is...@googlegroups.com.
To unsubscribe from this group, send email to 
open-iscsi+unsubscr...@googlegroups.com.
For more options, visit this group at 
http://groups.google.com/group/open-iscsi?hl=en.



Re: iscsi seems to work, but no device node?

2010-08-24 Thread Mike Christie

On 08/24/2010 11:43 AM, agshekeloh wrote:

Hi,

I'm running open-iscsi 2.0-871, on Ubuntu 10.04 amd64, straight from
the operating system package.  Mounting an iSCSI drive works perfectly
when booting from a hard drive, but not when booting diskless via
NFS.  The iSCSI server is OpenSolaris.

The diskless node can discover the iSCSI node and log in:

# iscsiadm -m session -P1
Target: iqn.1986-03.com.sun:02:2293f889-f1bb-e764-e45f-b7931c6c86a5
 Current Portal: XXX.XXX.64.168:3260,1
 Persistent Portal: XXX.XXX.64.168:3260,1
 **
 Interface:
 **
 Iface Name: default
 Iface Transport: tcp
 Iface Initiatorname: iqn.1993-08.org.debian:
01:cd00198dd976
 Iface IPaddress: XXX.XXX.199.20
 Iface HWaddress:
 Iface Netdev:
 SID: 1
 iSCSI Connection State: LOGGED IN
 iSCSI Session State: LOGGED_IN
 Internal iscsid Session State: NO CHANGE

I would expect to see /dev/sda at this point, but it doesn't appear.



What does iscsiadm -m session -P 3 print out? Could you also send the 
/var/log/messages output when you run the login command that fails to 
find disks.


Also if you do

echo - - - > /sys/class/scsi_host/hostX/scan

X is the host number you will see from the -P3 command.


Does that find disks (check /var/log/messages and /dev after the echo 
has completed).


--
You received this message because you are subscribed to the Google Groups 
"open-iscsi" group.
To post to this group, send email to open-is...@googlegroups.com.
To unsubscribe from this group, send email to 
open-iscsi+unsubscr...@googlegroups.com.
For more options, visit this group at 
http://groups.google.com/group/open-iscsi?hl=en.



iscsi seems to work, but no device node?

2010-08-24 Thread agshekeloh
Hi,

I'm running open-iscsi 2.0-871, on Ubuntu 10.04 amd64, straight from
the operating system package.  Mounting an iSCSI drive works perfectly
when booting from a hard drive, but not when booting diskless via
NFS.  The iSCSI server is OpenSolaris.

The diskless node can discover the iSCSI node and log in:

# iscsiadm -m session -P1
Target: iqn.1986-03.com.sun:02:2293f889-f1bb-e764-e45f-b7931c6c86a5
Current Portal: XXX.XXX.64.168:3260,1
Persistent Portal: XXX.XXX.64.168:3260,1
**
Interface:
**
Iface Name: default
Iface Transport: tcp
Iface Initiatorname: iqn.1993-08.org.debian:
01:cd00198dd976
Iface IPaddress: XXX.XXX.199.20
Iface HWaddress: 
Iface Netdev: 
SID: 1
iSCSI Connection State: LOGGED IN
iSCSI Session State: LOGGED_IN
Internal iscsid Session State: NO CHANGE

I would expect to see /dev/sda at this point, but it doesn't appear.
There's no nodes starting with "ip" in /dev/disk/by-path, either.

I ran iscsid -fd8 by hand and looked for errors, but it looked much
like the output from the working machine (at least to my untrained
eye).  Any suggestions on how to debug this further?

Thanks,
==ml

-- 
You received this message because you are subscribed to the Google Groups 
"open-iscsi" group.
To post to this group, send email to open-is...@googlegroups.com.
To unsubscribe from this group, send email to 
open-iscsi+unsubscr...@googlegroups.com.
For more options, visit this group at 
http://groups.google.com/group/open-iscsi?hl=en.



Confirm that bnx2i iSCSI offload works for RHEL 5.5 on DELL EqualLogic

2010-08-24 Thread Michael Wolf
Hi ALL,

I try to set up oracle cluster with four nodes (DELL R610) and
EqualLogig PS6000.
Servers has four Broadcom 5709C dual network cards, two intenal, two
PCI cards.
I`m trying to get the bnx2i driver to work using iSCSI offload for
four channels.
I can't get it to work, getting error
iscsiadm: Could not login to [iface: bnx2i.00:22:19:68:3a:04, target:
iqn.2001-05.com.equallogic:0-8a0906-76c630e04-e0492d35ddc63f-a1test,
portal: 10.10.77.10,3260):
iscsiadm: initiator reported error (4 - encountered connection
failure)

Showing nothing in /var/log/messages and no events on the EQ box.
DELL suport team is not shure if bnx2i driver will work with RHEL 5.5,
they ask me to reinstall RHEL 5.2, from my opinion is that not an
option.
I am using RHEL 5.5, oracle linux, kernel 2.6.18-194.11.1.0.1.el5 on
x86_64
iscsi-initiator-utils-6.2.0.871-0.16.el5
All firmware of the R610 is up to date

All interfaces are shown as bnx2i
[r...@ora-db-rac1-04 ~]# scsiadm -m iface
-bash: scsiadm: command not found
[r...@ora-db-rac1-04 ~]# iscsiadm -m iface
default tcp
iser iser
bnx2i.00:22:19:68:3a:04 bnx2i,00:22:19:68:3a:
04,10.10.77.69,,
bnx2i.00:22:19:68:3a:0a bnx2i,00:22:19:68:3a:0a,
10.10.77.68,,
bnx2i.00:00:00:00:00:00 bnx2i,
00:00:00:00:00:00,,,
bnx2i.00:22:19:68:3a:06 bnx2i,00:22:19:68:3a:
06,10.10.77.67,,
bnx2i.00:10:18:64:2e:5b bnx2i,00:10:18:64:2e:5b,
10.10.77.66,,
bnx2i.00:10:18:64:98:bb bnx2i,00:10:18:64:98:bb,
10.10.77.65,,
bnx2i.00:22:19:68:3a:08 bnx2i,00:22:19:68:3a:
08,10.10.77.64,,
bnx2i.00:10:18:64:98:b9 bnx2i,
00:10:18:64:98:b9,10.10.77.63,,

Only four of them are connected to the SAN. How to figure out which hw-
adress belongs to which interface?
Is bnx2i driver supported for RHEL5.5?
Is there a more detailed description for setting up such an
environment as the RedHat manuals?
I am searching for  the best way to figure out he reason for the
connect failures?

Thanks for any helping hand.

-- 
You received this message because you are subscribed to the Google Groups 
"open-iscsi" group.
To post to this group, send email to open-is...@googlegroups.com.
To unsubscribe from this group, send email to 
open-iscsi+unsubscr...@googlegroups.com.
For more options, visit this group at 
http://groups.google.com/group/open-iscsi?hl=en.



Re: detected conn error (1011)

2010-08-24 Thread Goncalo Gomes
Hi,

On Fri, 2010-08-06 at 15:57 +0100, Hannes Reinecke wrote: 
> Mike Christie wrote:
> > ccing Hannes from suse, because this looks like a SLES only bug.
> > 
> > Hey Hannes,
> > 
> > The user is using Linux 2.6.27 x86 based on SLES + Xen 3.4 (as dom0)
> > running a couple of RHEL 5.5 VMs. The underlying storage for these VMs
> > is iSCSI based via open-iscsi 2.0.870-26.6.1 and a DELL equallogic array.
> > 
> > 
> > On 08/05/2010 02:21 PM, Goncalo Gomes wrote:
> >> I've copied both the messages file from the host goncalog140 and the
> >> patched libiscsi.c. FWIW, I've also included the iscsid.conf. Find these
> >> files in the link below:
> >>
> >> http://promisc.org/iscsi/
> >>
> > 
> > It looks like this chunk from libiscsi.c:iscsi_queuecommand:
> > 
> > case ISCSI_STATE_FAILED:
> > reason = FAILURE_SESSION_FAILED;
> > sc->result = DID_TRANSPORT_DISRUPTED << 16;
> > break;
> > 
> > is causing IO errors.
> > 
> > You want to use something like DID_IMM_RETRY because it can be a long
> > time between the time the kernel marks the state as ISCSI_STATE_FAILED
> > until we start recovery and properly get all the device queues blocked,
> > so we can exhaust all the retries if we use DID_TRANSPORT_DISRUPTED.
> Yeah, I noticed.
> But the problem is that multipathing will stall during this time,
> ie no failover will occur and I/O will stall. Using DID_TRANSPORT_DISRUPTED
> will circumvent this and we can failover immediately.
> 
> Sadly I got additional bugreports about this so I think I'll have
> to revert it.

I applied and tested the changes Mike Christie suggests. After the LUN
is rebalanced within the array I no longer see the IO errors and it
appears the setup is now resilient to the equallogic LUN failover
process.

I'm attaching the log from the dmesg merely for sanity check purposes,
if anyone cares to take a look?

> I have put some test kernels at
> 
> http://beta.suse.com/private/hare/sles11/iscsi

Do the test kernels in the url above contain the change of
DID_TRANSPORT_DISRUPTED to DID_DIMM_RETRY or is there more to it than
simply changing the result code? If the latter, would you be able to
upload the source rpms or a unified patch containing the changes you are
are staging? I'm looking for a more pallatable way to test them, given I
have no SLES box lying around, but will install one if needs be.

Thanks,
-Goncalo.

-- 
You received this message because you are subscribed to the Google Groups 
"open-iscsi" group.
To post to this group, send email to open-is...@googlegroups.com.
To unsubscribe from this group, send email to 
open-iscsi+unsubscr...@googlegroups.com.
For more options, visit this group at 
http://groups.google.com/group/open-iscsi?hl=en.

device-mapper: multipath: version 1.0.5 loaded
device-mapper: multipath round-robin: version 1.0.0 loaded
device-mapper: table: 251:1: multipath: error getting device
device-mapper: ioctl: error adding target to table
device-mapper: table: 251:1: multipath: error getting device
device-mapper: ioctl: error adding target to table
Citrix Systems, Inc. -- Private Release Kernel
Private File Disclaimer The private files provided to you contain a preliminary 
code fix. These private files have been created and distributed to you to 
address your specific issue and provide Citrix with the feedback that your 
issue has been resolved or to provide further debugging information. These 
private files have had minimal in-house testing with no regression testing and 
may contain defects.  These private file(s) will only be supported until an 
official Hotfix has been provided or one is publicly available from the Citrix 
web site. Any private files that are provided to you are intended only for the 
use of the individual or entity to which this is addressed and distribution of 
these files or utilities is prohibited. CITRIX MAKES NO REPRESENTATIONS OR 
WARRANTIES OF NONINFRINGEMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR 
PURPOSE WITH RESPECT TO THE PRIVATE FILES.  THE PRIVATE FILES ARE DELIVERED ON 
AN "AS IS" BASIS. YOU SHALL HAVE THE SOLE RESPONSIBILITY FOR ADEQUATE 
PROTECTION AND BACK-UP OF AN<6>Loading iSCSI transport class v2.0-870.
iscsi: registered transport (tcp)
scsi6 : iSCSI Initiator over TCP/IP
 connection1:0: detected conn error (1011)
scsi 6:0:0:0: Direct-Access EQLOGIC  100E-00  4.3  PQ: 0 ANSI: 5
sd 6:0:0:0: [sdb] 209725440 512-byte hardware sectors: (107 GB/100 GiB)
sd 6:0:0:0: [sdb] Write Protect is off
sd 6:0:0:0: [sdb] Mode Sense: ad 00 00 00
sd 6:0:0:0: [sdb] Write cache: disabled, read cache: enabled, doesn't support 
DPO or FUA
sd 6:0:0:0: [sdb] 209725440 512-byte hardware sectors: (107 GB/100 GiB)
sd 6:0:0:0: [sdb] Write Protect is off
sd 6:0:0:0: [sdb] Mode Sense: ad 00 00 00
sd 6:0:0:0: [sdb] Write cache: disabled, read cache: enabled, doesn't support 
DPO or FUA
 sdb: sdb1
sd 6:0:0:0: [sdb] Attached SCSI disk
sd 6:0:0:0: Attached scsi generic sg1 type 0
tap_backend_changed: