OK Today I played with a J4400 connected to a Txxx server running S10 10/09

First off read the release notes I spent about 4 hours pulling my hair out as I 
could not get stmsboot to work until we read in the release notes that 500GB 
SATA drives do not work!!!

Initial Setup:
A pair of dual port SAS controllers (c4 and c5)
A J4400 with 6x 1TB SATA disks

The J440 had two controllers and these where connected to one SAS card 
(physical controller c4)

Test 1:

First a reboot -- -r

format shows 12 disks on c4 (each disk having two paths). If you picked the 
same disk via both paths ZFS stopped you doing stupid things by knowing the 
disk was already in use.

Test 2:

run stmsboot -e

format now shows six disk on controller c6, a new "virtual controller" The two 
internal disks are also now on c6 and stmsboot has done the right stuff with 
the rpool, so I would guess you could multi-path at a later date if you don't 
want to fist off, but I did not test this.

stmsboot -L only showed the two internal disk not the six in the J4400 strange, 
but we pressed on.

Test 3:

I created a zpool (two disks mirrored) using two of the new devices on c6.

I created some I/O load

I then unplugged one of the cables from the SAS card (physical c4).

Result: Nothing everything just keeps working - cool stuff!

Test 4:

I plugged the unplugged cable into the other controller (physical c5)

Result: Nothing everything just keeps working - cool stuff!

Test 5:

Being bold I then unplugged the remaining cable from the physical c4 controller

Result: Nothing everything just keeps working - cool stuff!

So I had gone from dual pathed, on a single controller (c4) to single pathed, 
on a different controller (c5).


Test 6:

I added the other four drives to the zpool (plain old zfs stuff - a bit boring).


Test 7:

I plugged in four more disks.

Result: Their mulipathed devices just showed up in format, I added them to the 
pool and also added them as spares all the while the I/O load is happening. No 
noticable stops or glitches.

Conclusion:

If you RTFM first then stmsboot does everything it is documented to do. You 
don't need to play with cfgadm or anything like that, just as I said orginally 
(below). The multi-pathing stuff is easy to set up and even a very rusty admin. 
like me found it very easy.

Note: There may be patches for the 500GB SATA disks I don'y know, fortunatly 
that's not what I've sold - Phew!!

TTFN
Trevor








________________________________
From: zfs-discuss-boun...@opensolaris.org [zfs-discuss-boun...@opensolaris.org] 
On Behalf Of Trevor Pretty [trevor_pre...@eagle.co.nz]
Sent: Monday, 30 November 2009 2:48 p.m.
To: Karl Katzke
Cc: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] Opensolaris with J4400 - Experiences

Karl

Don't you just use stmsboot?

http://docs.sun.com/source/820-3223-14/SASMultipath.html#50511899_pgfId-1046940

Bruno

Next week I'm playing with a M3000 and a J4200 in the local NZ distributor's  
lab. I had planned to just use the latest version of S10, but if I get the time 
I might play with OpenSolaris as well, but I don't think there is anything 
radically different between the two here.

>From what I've read in preparation (and I stand to be corrected):



    * Will i be able to achieve multipath support, if i connect the
      J4400 to 2 LSI HBA in one server, with SATA disks, or this is only
      possible with SAS disks? This server will have OpenSolaris (any
      release i think) .

Disk type does not matter (see link above).

    * The CAM ( StorageTek Common Array Manager ), its only for hardware
      management of the JBOD, leaving
      disk/volumes/zpools/luns/whatever_name management up to the server
      operating system , correct ?

That is my understanding see:- http://docs.sun.com/source/820-3765-11/

    * Can i put some readzillas/writezillas in the j4400 along with sata
      disks, and if so will i have any benefit  , or should i place
      those *zillas directly into the servers disk tray?

On the Unified Storage products they go in both. Readzilla in the server 
Logzillas in the J4400. This is quite logical if you want to move the array 
between hosts all the data needs to be in the array. Read data can always be 
re-created so therefore the closer to the CPU the better. See: 
http://catalog.sun.com/

    * Does any one has experiences with those jbods? If so, are they in
      general solid/reliable ?

No: But, get a support contract!

    * The server will probably be a Sun x44xx series, with 32Gb ram, but
      for the best possible performance, should i invest in more and
      more spindles, or a couple less spindles and buy some readzillas?
      This system will be mainly used to export some volumes over ISCSI
      to a windows 2003 fileserver, and to hold some NFS shares.

Check Brendon Gregg's blogs *I think* he has done some work here from memory.







Karl Katzke wrote:

Bruno -

Sorry, I don't have experience with OpenSolaris, but I *do* have experience 
running a J4400 with Solaris 10u8.

First off, you need a LSI HBA for the Multipath support. It won't work with any 
others as far as I know.

I ran into problems with the multipath support because it wouldn't allow me to 
manage the disks with cfgadm and got very confused when I'd do something as 
silly as replace a disk, causing the disk's GUID (and therefor address under 
the virtual multipath controller) to change. My take-away was that Solaris 10u8 
multipath support is not ready for production environments as there are 
limited-to-no administration tools. This may have been fixed in recent builds 
of Nevada. (See a thread that started around 03Nov09 for my experiences with 
MPxIO.)

At the moment, I have the J4400 split between the two controllers and simply 
have even numbered disks on one, and odd numbered disks on the other. Both 
controllers can *see* all the disks.

You are correct about the CAM software. It also updates the firmware, though, 
since us commoners don't seemingly have access to the serial management ports 
on the J4400.

I can't speak to locating the drives -- that would be something you'd have to 
test. I have found increases in performance on my faster and more random array; 
others have found exactly the opposite.

My configuration is as follows;
x4250
- rpool - 2x 146 gb 10k SAS
- 'hot' pool - 10x 300gb 10k SAS + 2x 32gb ZIL
j4400
- 'cold' pool - 12x 1tb 7200rpm SATA ... testing adding 2x 146gb SAS in the 
x4250, but haven't benchmarked yet.

Performance on the J4400 was disappointing with just one controller to 12 disks 
in one RAIDZ2 and no ZIL. However, I do not know if the bottleneck was at the 
disk, controller, backplane, or software level... I'm too close to my deadline 
to do much besides randomly shotgunning different configs to see what works 
best!

-K


Karl Katzke
Systems Analyst II
TAMU - RGS





On 11/25/2009 at 11:13 AM, in message 
<4b0d65d6.4020...@epinfante.com><mailto:4b0d65d6.4020...@epinfante.com>, Bruno


Sousa <bso...@epinfante.com><mailto:bso...@epinfante.com> wrote:


Hello !

I'm currently using a X2200 with a LSI HBA connected to a Supermicro
JBOD chassis, however i want to have more redundancy in the JBOD.
So i have looked into to market, and into to the wallet, and i think
that the Sun J4400 suits nicely to my goals. However i have some
concerns and if anyone can give some suggestions i would trully appreciate.
And now for my questions :

    * Will i be able to achieve multipath support, if i connect the
      J4400 to 2 LSI HBA in one server, with SATA disks, or this is only
      possible with SAS disks? This server will have OpenSolaris (any
      release i think) .
    * The CAM ( StorageTek Common Array Manager ), its only for hardware
      management of the JBOD, leaving
      disk/volumes/zpools/luns/whatever_name management up to the server
      operating system , correct ?
    * Can i put some readzillas/writezillas in the j4400 along with sata
      disks, and if so will i have any benefit  , or should i place
      those *zillas directly into the servers disk tray?
    * Does any one has experiences with those jbods? If so, are they in
      general solid/reliable ?
    * The server will probably be a Sun x44xx series, with 32Gb ram, but
      for the best possible performance, should i invest in more and
      more spindles, or a couple less spindles and buy some readzillas?
      This system will be mainly used to export some volumes over ISCSI
      to a windows 2003 fileserver, and to hold some NFS shares.


Thank you for all your time,
Bruno



_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org<mailto:zfs-discuss@opensolaris.org>
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

www.eagle.co.nz<http://www.eagle.co.nz/>

This email is confidential and may be legally privileged. If received in error 
please destroy and immediately notify us.

===
www.eagle.co.nz 

This email is confidential and may be legally privileged.
If received in error please destroy and immediately notify us.
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to