Re: [ceph-users] Ceph with VMWare / XenServer

2014-05-16 Thread Andrei Mikhailovsky
Uwe, 

could you please help me a bit with configuring multipathing on two different 
storage servers and connecting it to xenserver. 

I am looking at the multipathing howto and it tells me that for multipathing to 
work the iscsi querry from the target server should return two paths. However, 
if you have two separate servers with tgt installed, each one would only return 
a single path. 

I've configured two servers (tgt1 and tgt2) with tgt, each pointing to the same 
rbd image. the iscsi config files are identical. One server is using 
192.168.170.200 ip, the second one uses 192.168.171.200. When doing a query, 
the tgt1, it returns: 


192.168.170.200:3260,1 
iqn.2014-04.iscsi-ibstorage.arhont.com:xenserver-iscsi-export-10TB-1 

and tgt2 returns: 

192.168.171.200:3260,1 
iqn.2014-04.iscsi-ibstorage.arhont.com:xenserver-iscsi-export-10TB-1 


According to the documentation, each server should return both paths, like 
this: 
192.168.170.200:3260,1 
iqn.2014-04.iscsi-ibstorage.arhont.com:xenserver-iscsi-export-10TB-1 
192.168.171.200:3260,1 
iqn.2014-04.iscsi-ibstorage.arhont.com:xenserver-iscsi-export-10TB-1 


Is there a manual way of configuring multipathing? Or have I not created the 
tgt configs correctly? 

Cheers 

Andrei 

- Original Message -

From: "Uwe Grohnwaldt"  
To: ceph-users@lists.ceph.com 
Sent: Monday, 12 May, 2014 12:57:48 PM 
Subject: Re: [ceph-users] Ceph with VMWare / XenServer 

Hi, 

at the moment we are using tgt with RBD backend compiled from source on Ubuntu 
12.04 and 14.04 LTS. We have two machines within two ip-ranges (e.g. 
192.168.1.0/24 and 192.168.2.0/24). One machine in 192.168.1.0/24 and one 
machine in 192.168.2.0/24. The config for tgt is the same on both machines, 
they export the same rbd. This works well for XenServer. 

For VMWare you have to disable VAAI to use it with tgt 
(http://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1033665)
 If you don't disable it, ESXi becomes very slow and unresponsive. 

I think the problem is the iSCSI Write Same Support but I haven't tried which 
of the settings of VAAI is responsible for this behavior. 

Mit freundlichen Grüßen / Best Regards, 
-- 
Consultant 
Dipl.-Inf. Uwe Grohnwaldt 
Gutleutstr. 351 
60327 Frankfurt a. M. 

eMail: u...@grohnwaldt.eu 
Telefon: +49-69-34878906 
Mobil: +49-172-3209285 
Fax: +49-69-348789069 

- Original Message - 
> From: "Andrei Mikhailovsky"  
> To: ceph-users@lists.ceph.com 
> Sent: Montag, 12. Mai 2014 12:00:48 
> Subject: [ceph-users] Ceph with VMWare / XenServer 
> 
> 
> 
> Hello guys, 
> 
> I am currently running a ceph cluster for running vms with qemu + 
> rbd. It works pretty well and provides a good degree of failover. I 
> am able to run maintenance tasks on the ceph nodes without 
> interrupting vms IO. 
> 
> I would like to do the same with VMWare / XenServer hypervisors, but 
> I am not really sure how to achieve this. Initially I thought of 
> using iscsi multipathing, however, as it turns out, multipathing is 
> more for load balancing and nic/switch failure. It does not allow me 
> to perform maintenance on the iscsi target without interrupting 
> service to vms. 
> 
> Has anyone done either a PoC or better a production environment where 
> they've used ceph as a backend storage with vmware / xenserver? The 
> important element for me is to have the ability of performing 
> maintenance tasks and resilience to failovers without interrupting 
> IO to vms. Are there any recommendations or howtos on how this could 
> be achieved? 
> 
> Many thanks 
> 
> Andrei 
> 
> 
> ___ 
> ceph-users mailing list 
> ceph-users@lists.ceph.com 
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
> 
___ 
ceph-users mailing list 
ceph-users@lists.ceph.com 
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Ceph with VMWare / XenServer

2014-05-13 Thread Andrei Mikhailovsky
Uwe, do you mind sharing your storage and xenserver iscsi config files? 

Also, what is your performance like? 

Thanks 

- Original Message -

From: "Uwe Grohnwaldt"  
To: ceph-users@lists.ceph.com 
Sent: Monday, 12 May, 2014 2:45:43 PM 
Subject: Re: [ceph-users] Ceph with VMWare / XenServer 

Hi, 

yes, we use it in production. I can stop/kill the tgt on one server and 
XenServer goes to the second one. We enabled multipathing in xenserver. In our 
setup we haven't multiple ip-ranges so we scan/login the second target on 
xenserverstartup with iscsiadm in rc.local. 

Thats based on history - we used Dell Equallogic before ceph came in and there 
was no need to use multipathing (only LACP-channels). No we enabled 
multipathing and use tgt, but without diffent ip-ranges. 

Mit freundlichen Grüßen / Best Regards, 
-- 
Consultant 
Dipl.-Inf. Uwe Grohnwaldt 
Gutleutstr. 351 
60327 Frankfurt a. M. 

eMail: u...@grohnwaldt.eu 
Telefon: +49-69-34878906 
Mobil: +49-172-3209285 
Fax: +49-69-348789069 

- Original Message - 
> From: "Andrei Mikhailovsky"  
> To: "Uwe Grohnwaldt"  
> Cc: ceph-users@lists.ceph.com 
> Sent: Montag, 12. Mai 2014 14:48:58 
> Subject: Re: [ceph-users] Ceph with VMWare / XenServer 
> 
> 
> Uwe, thanks for your quick reply. 
> 
> Do you run the Xenserver setup on production env and have you tried 
> to test some failover scenarios to see if the xenserver guest vms 
> are working during the failover of storage servers? 
> 
> Also, how did you set up the xenserver iscsi? Have you used the 
> multipath option to set up the LUNs? 
> 
> Cheers 
> 
> 
> 
> 
> - Original Message - 
> 
> From: "Uwe Grohnwaldt"  
> To: ceph-users@lists.ceph.com 
> Sent: Monday, 12 May, 2014 12:57:48 PM 
> Subject: Re: [ceph-users] Ceph with VMWare / XenServer 
> 
> Hi, 
> 
> at the moment we are using tgt with RBD backend compiled from source 
> on Ubuntu 12.04 and 14.04 LTS. We have two machines within two 
> ip-ranges (e.g. 192.168.1.0/24 and 192.168.2.0/24). One machine in 
> 192.168.1.0/24 and one machine in 192.168.2.0/24. The config for tgt 
> is the same on both machines, they export the same rbd. This works 
> well for XenServer. 
> 
> For VMWare you have to disable VAAI to use it with tgt 
> (http://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1033665)
>  
> If you don't disable it, ESXi becomes very slow and unresponsive. 
> 
> I think the problem is the iSCSI Write Same Support but I haven't 
> tried which of the settings of VAAI is responsible for this 
> behavior. 
> 
> Mit freundlichen Grüßen / Best Regards, 
> -- 
> Consultant 
> Dipl.-Inf. Uwe Grohnwaldt 
> Gutleutstr. 351 
> 60327 Frankfurt a. M. 
> 
> eMail: u...@grohnwaldt.eu 
> Telefon: +49-69-34878906 
> Mobil: +49-172-3209285 
> Fax: +49-69-348789069 
> 
> - Original Message - 
> > From: "Andrei Mikhailovsky"  
> > To: ceph-users@lists.ceph.com 
> > Sent: Montag, 12. Mai 2014 12:00:48 
> > Subject: [ceph-users] Ceph with VMWare / XenServer 
> > 
> > 
> > 
> > Hello guys, 
> > 
> > I am currently running a ceph cluster for running vms with qemu + 
> > rbd. It works pretty well and provides a good degree of failover. I 
> > am able to run maintenance tasks on the ceph nodes without 
> > interrupting vms IO. 
> > 
> > I would like to do the same with VMWare / XenServer hypervisors, 
> > but 
> > I am not really sure how to achieve this. Initially I thought of 
> > using iscsi multipathing, however, as it turns out, multipathing is 
> > more for load balancing and nic/switch failure. It does not allow 
> > me 
> > to perform maintenance on the iscsi target without interrupting 
> > service to vms. 
> > 
> > Has anyone done either a PoC or better a production environment 
> > where 
> > they've used ceph as a backend storage with vmware / xenserver? The 
> > important element for me is to have the ability of performing 
> > maintenance tasks and resilience to failovers without interrupting 
> > IO to vms. Are there any recommendations or howtos on how this 
> > could 
> > be achieved? 
> > 
> > Many thanks 
> > 
> > Andrei 
> > 
> > 
> > ___ 
> > ceph-users mailing list 
> > ceph-users@lists.ceph.com 
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
> > 
> ___ 
> ceph-users mailing list 
> ceph-users@lists.ceph.com 
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
> 
> 
___ 
ceph-users mailing list 
ceph-users@lists.ceph.com 
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Ceph with VMWare / XenServer

2014-05-13 Thread Gilles Mocellin

Le 12/05/2014 15:45, Uwe Grohnwaldt a écrit :

Hi,

yes, we use it in production. I can stop/kill the tgt on one server and 
XenServer goes to the second one. We enabled multipathing in xenserver. In our 
setup we haven't multiple ip-ranges so we scan/login the second target on 
xenserverstartup with iscsiadm in rc.local.

Thats based on history - we used Dell Equallogic before ceph came in and there 
was no need to use multipathing (only LACP-channels). No we enabled 
multipathing and use tgt, but without diffent ip-ranges.



So you use multipathing in failover mode, that's certainly why it works 
without state sharing between the tgtd servers.

Still, I think you need to deactivate all sort of caching server side.
IO must be committed to ceph when the iSCSI initiator think it is.

What are the multipath parameters in XenServer (timeout, retry...) ?


___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Ceph with VMWare / XenServer

2014-05-12 Thread Uwe Grohnwaldt
We are using a switchstack of Juniper EX4200 and EX3200. Cisco should work, 
too. Anotheroption is failover bonding (but multipathing with different ips is 
better)

Mit freundlichen Grüßen / Best Regards,
--
Consultant
Dipl.-Inf. Uwe Grohnwaldt
Gutleutstr. 351
60327 Frankfurt a. M.

eMail: u...@grohnwaldt.eu
Telefon: +49-69-34878906
Mobil: +49-172-3209285
Fax: +49-69-348789069

- Original Message -
> From: "Leen Besselink" 
> To: ceph-users@lists.ceph.com
> Cc: "Uwe Grohnwaldt" 
> Sent: Montag, 12. Mai 2014 19:01:46
> Subject: Re: [ceph-users] Ceph with VMWare / XenServer
> 
> On Mon, May 12, 2014 at 03:45:43PM +0200, Uwe Grohnwaldt wrote:
> > Hi,
> > 
> > yes, we use it in production. I can stop/kill the tgt on one server
> > and XenServer goes to the second one. We enabled multipathing in
> > xenserver. In our setup we haven't multiple ip-ranges so we
> > scan/login the second target on xenserverstartup with iscsiadm in
> > rc.local.
> > 
> > Thats based on history - we used Dell Equallogic before ceph came
> > in and there was no need to use multipathing (only LACP-channels).
> > No we enabled multipathing and use tgt, but without diffent
> > ip-ranges.
> > 
> 
> I assume you connected the machines to the same switch ? As normal
> LACP don't work with multiple switches.
> 
> Is that correct ?
> 
> It wasn't that I needed different ip-ranges in my setup, it just
> makes it simpler/predictable.
> 
> > Mit freundlichen Grüßen / Best Regards,
> > --
> > Consultant
> > Dipl.-Inf. Uwe Grohnwaldt
> > Gutleutstr. 351
> > 60327 Frankfurt a. M.
> > 
> > eMail: u...@grohnwaldt.eu
> > Telefon: +49-69-34878906
> > Mobil: +49-172-3209285
> > Fax: +49-69-348789069
> > 
> > - Original Message -
> > > From: "Andrei Mikhailovsky" 
> > > To: "Uwe Grohnwaldt" 
> > > Cc: ceph-users@lists.ceph.com
> > > Sent: Montag, 12. Mai 2014 14:48:58
> > > Subject: Re: [ceph-users] Ceph with VMWare / XenServer
> > > 
> > > 
> > > Uwe, thanks for your quick reply.
> > > 
> > > Do you run the Xenserver setup on production env and have you
> > > tried
> > > to test some failover scenarios to see if the xenserver guest vms
> > > are working during the failover of storage servers?
> > > 
> > > Also, how did you set up the xenserver iscsi? Have you used the
> > > multipath option to set up the LUNs?
> > > 
> > > Cheers
> > > 
> > > 
> > > 
> > > 
> > > - Original Message -
> > > 
> > > From: "Uwe Grohnwaldt" 
> > > To: ceph-users@lists.ceph.com
> > > Sent: Monday, 12 May, 2014 12:57:48 PM
> > > Subject: Re: [ceph-users] Ceph with VMWare / XenServer
> > > 
> > > Hi,
> > > 
> > > at the moment we are using tgt with RBD backend compiled from
> > > source
> > > on Ubuntu 12.04 and 14.04 LTS. We have two machines within two
> > > ip-ranges (e.g. 192.168.1.0/24 and 192.168.2.0/24). One machine
> > > in
> > > 192.168.1.0/24 and one machine in 192.168.2.0/24. The config for
> > > tgt
> > > is the same on both machines, they export the same rbd. This
> > > works
> > > well for XenServer.
> > > 
> > > For VMWare you have to disable VAAI to use it with tgt
> > > (http://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1033665)
> > > If you don't disable it, ESXi becomes very slow and unresponsive.
> > > 
> > > I think the problem is the iSCSI Write Same Support but I haven't
> > > tried which of the settings of VAAI is responsible for this
> > > behavior.
> > > 
> > > Mit freundlichen Grüßen / Best Regards,
> > > --
> > > Consultant
> > > Dipl.-Inf. Uwe Grohnwaldt
> > > Gutleutstr. 351
> > > 60327 Frankfurt a. M.
> > > 
> > > eMail: u...@grohnwaldt.eu
> > > Telefon: +49-69-34878906
> > > Mobil: +49-172-3209285
> > > Fax: +49-69-348789069
> > > 
> > > - Original Message -
> > > > From: "Andrei Mikhailovsky" 
> > > > To: ceph-users@lists.ceph.com
> > > > Sent: Montag, 12. Mai 2014 12:00:48
> > > > Subject: [ceph-users] Ceph with VMWare / XenServer
> > > > 
> > > > 
> > > > 
> > > > Hello guys,
> > > > 
> > > > I am

Re: [ceph-users] Ceph with VMWare / XenServer

2014-05-12 Thread Leen Besselink
On Mon, May 12, 2014 at 07:01:46PM +0200, Leen Besselink wrote:
> On Mon, May 12, 2014 at 03:45:43PM +0200, Uwe Grohnwaldt wrote:
> > Hi,
> > 
> > yes, we use it in production. I can stop/kill the tgt on one server and 
> > XenServer goes to the second one. We enabled multipathing in xenserver. In 
> > our setup we haven't multiple ip-ranges so we scan/login the second target 
> > on xenserverstartup with iscsiadm in rc.local.
> > 
> > Thats based on history - we used Dell Equallogic before ceph came in and 
> > there was no need to use multipathing (only LACP-channels). No we enabled 
> > multipathing and use tgt, but without diffent ip-ranges.
> > 
> 
> I assume you connected the machines to the same switch ? As normal LACP don't 
> work with multiple switches.
> 
> Is that correct ?
> 

Or maybe you used a stack or you have Cisco switches with vPC ?

> It wasn't that I needed different ip-ranges in my setup, it just makes it 
> simpler/predictable.
> 
> > Mit freundlichen Grüßen / Best Regards,
> > --
> > Consultant
> > Dipl.-Inf. Uwe Grohnwaldt
> > Gutleutstr. 351
> > 60327 Frankfurt a. M.
> > 
> > eMail: u...@grohnwaldt.eu
> > Telefon: +49-69-34878906
> > Mobil: +49-172-3209285
> > Fax: +49-69-348789069
> > 
> > - Original Message -
> > > From: "Andrei Mikhailovsky" 
> > > To: "Uwe Grohnwaldt" 
> > > Cc: ceph-users@lists.ceph.com
> > > Sent: Montag, 12. Mai 2014 14:48:58
> > > Subject: Re: [ceph-users] Ceph with VMWare / XenServer
> > > 
> > > 
> > > Uwe, thanks for your quick reply.
> > > 
> > > Do you run the Xenserver setup on production env and have you tried
> > > to test some failover scenarios to see if the xenserver guest vms
> > > are working during the failover of storage servers?
> > > 
> > > Also, how did you set up the xenserver iscsi? Have you used the
> > > multipath option to set up the LUNs?
> > > 
> > > Cheers
> > > 
> > > 
> > > 
> > > 
> > > - Original Message -
> > > 
> > > From: "Uwe Grohnwaldt" 
> > > To: ceph-users@lists.ceph.com
> > > Sent: Monday, 12 May, 2014 12:57:48 PM
> > > Subject: Re: [ceph-users] Ceph with VMWare / XenServer
> > > 
> > > Hi,
> > > 
> > > at the moment we are using tgt with RBD backend compiled from source
> > > on Ubuntu 12.04 and 14.04 LTS. We have two machines within two
> > > ip-ranges (e.g. 192.168.1.0/24 and 192.168.2.0/24). One machine in
> > > 192.168.1.0/24 and one machine in 192.168.2.0/24. The config for tgt
> > > is the same on both machines, they export the same rbd. This works
> > > well for XenServer.
> > > 
> > > For VMWare you have to disable VAAI to use it with tgt
> > > (http://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1033665)
> > > If you don't disable it, ESXi becomes very slow and unresponsive.
> > > 
> > > I think the problem is the iSCSI Write Same Support but I haven't
> > > tried which of the settings of VAAI is responsible for this
> > > behavior.
> > > 
> > > Mit freundlichen Grüßen / Best Regards,
> > > --
> > > Consultant
> > > Dipl.-Inf. Uwe Grohnwaldt
> > > Gutleutstr. 351
> > > 60327 Frankfurt a. M.
> > > 
> > > eMail: u...@grohnwaldt.eu
> > > Telefon: +49-69-34878906
> > > Mobil: +49-172-3209285
> > > Fax: +49-69-348789069
> > > 
> > > - Original Message -
> > > > From: "Andrei Mikhailovsky" 
> > > > To: ceph-users@lists.ceph.com
> > > > Sent: Montag, 12. Mai 2014 12:00:48
> > > > Subject: [ceph-users] Ceph with VMWare / XenServer
> > > > 
> > > > 
> > > > 
> > > > Hello guys,
> > > > 
> > > > I am currently running a ceph cluster for running vms with qemu +
> > > > rbd. It works pretty well and provides a good degree of failover. I
> > > > am able to run maintenance tasks on the ceph nodes without
> > > > interrupting vms IO.
> > > > 
> > > > I would like to do the same with VMWare / XenServer hypervisors,
> > > > but
> > > > I am not really sure how to achieve this. Initially I thought of
> > > > using iscsi multipathing, however, as it turns out, multipathing is
> >

Re: [ceph-users] Ceph with VMWare / XenServer

2014-05-12 Thread Leen Besselink
On Mon, May 12, 2014 at 03:45:43PM +0200, Uwe Grohnwaldt wrote:
> Hi,
> 
> yes, we use it in production. I can stop/kill the tgt on one server and 
> XenServer goes to the second one. We enabled multipathing in xenserver. In 
> our setup we haven't multiple ip-ranges so we scan/login the second target on 
> xenserverstartup with iscsiadm in rc.local.
> 
> Thats based on history - we used Dell Equallogic before ceph came in and 
> there was no need to use multipathing (only LACP-channels). No we enabled 
> multipathing and use tgt, but without diffent ip-ranges.
> 

I assume you connected the machines to the same switch ? As normal LACP don't 
work with multiple switches.

Is that correct ?

It wasn't that I needed different ip-ranges in my setup, it just makes it 
simpler/predictable.

> Mit freundlichen Grüßen / Best Regards,
> --
> Consultant
> Dipl.-Inf. Uwe Grohnwaldt
> Gutleutstr. 351
> 60327 Frankfurt a. M.
> 
> eMail: u...@grohnwaldt.eu
> Telefon: +49-69-34878906
> Mobil: +49-172-3209285
> Fax: +49-69-348789069
> 
> - Original Message -
> > From: "Andrei Mikhailovsky" 
> > To: "Uwe Grohnwaldt" 
> > Cc: ceph-users@lists.ceph.com
> > Sent: Montag, 12. Mai 2014 14:48:58
> > Subject: Re: [ceph-users] Ceph with VMWare / XenServer
> > 
> > 
> > Uwe, thanks for your quick reply.
> > 
> > Do you run the Xenserver setup on production env and have you tried
> > to test some failover scenarios to see if the xenserver guest vms
> > are working during the failover of storage servers?
> > 
> > Also, how did you set up the xenserver iscsi? Have you used the
> > multipath option to set up the LUNs?
> > 
> > Cheers
> > 
> > 
> > 
> > 
> > - Original Message -
> > 
> > From: "Uwe Grohnwaldt" 
> > To: ceph-users@lists.ceph.com
> > Sent: Monday, 12 May, 2014 12:57:48 PM
> > Subject: Re: [ceph-users] Ceph with VMWare / XenServer
> > 
> > Hi,
> > 
> > at the moment we are using tgt with RBD backend compiled from source
> > on Ubuntu 12.04 and 14.04 LTS. We have two machines within two
> > ip-ranges (e.g. 192.168.1.0/24 and 192.168.2.0/24). One machine in
> > 192.168.1.0/24 and one machine in 192.168.2.0/24. The config for tgt
> > is the same on both machines, they export the same rbd. This works
> > well for XenServer.
> > 
> > For VMWare you have to disable VAAI to use it with tgt
> > (http://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1033665)
> > If you don't disable it, ESXi becomes very slow and unresponsive.
> > 
> > I think the problem is the iSCSI Write Same Support but I haven't
> > tried which of the settings of VAAI is responsible for this
> > behavior.
> > 
> > Mit freundlichen Grüßen / Best Regards,
> > --
> > Consultant
> > Dipl.-Inf. Uwe Grohnwaldt
> > Gutleutstr. 351
> > 60327 Frankfurt a. M.
> > 
> > eMail: u...@grohnwaldt.eu
> > Telefon: +49-69-34878906
> > Mobil: +49-172-3209285
> > Fax: +49-69-348789069
> > 
> > - Original Message -
> > > From: "Andrei Mikhailovsky" 
> > > To: ceph-users@lists.ceph.com
> > > Sent: Montag, 12. Mai 2014 12:00:48
> > > Subject: [ceph-users] Ceph with VMWare / XenServer
> > > 
> > > 
> > > 
> > > Hello guys,
> > > 
> > > I am currently running a ceph cluster for running vms with qemu +
> > > rbd. It works pretty well and provides a good degree of failover. I
> > > am able to run maintenance tasks on the ceph nodes without
> > > interrupting vms IO.
> > > 
> > > I would like to do the same with VMWare / XenServer hypervisors,
> > > but
> > > I am not really sure how to achieve this. Initially I thought of
> > > using iscsi multipathing, however, as it turns out, multipathing is
> > > more for load balancing and nic/switch failure. It does not allow
> > > me
> > > to perform maintenance on the iscsi target without interrupting
> > > service to vms.
> > > 
> > > Has anyone done either a PoC or better a production environment
> > > where
> > > they've used ceph as a backend storage with vmware / xenserver? The
> > > important element for me is to have the ability of performing
> > > maintenance tasks and resilience to failovers without interrupting
> > > IO to vms. Are there any recommendations or howtos on how this
> > > could
> > > be achieved?
> > > 
> > > Many thanks
> > > 
> > > Andrei
> > > 
> > > 
> > > ___
> > > ceph-users mailing list
> > > ceph-users@lists.ceph.com
> > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > > 
> > ___
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > 
> > 
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Ceph with VMWare / XenServer

2014-05-12 Thread Uwe Grohnwaldt
Hi,

yes, we use it in production. I can stop/kill the tgt on one server and 
XenServer goes to the second one. We enabled multipathing in xenserver. In our 
setup we haven't multiple ip-ranges so we scan/login the second target on 
xenserverstartup with iscsiadm in rc.local.

Thats based on history - we used Dell Equallogic before ceph came in and there 
was no need to use multipathing (only LACP-channels). No we enabled 
multipathing and use tgt, but without diffent ip-ranges.

Mit freundlichen Grüßen / Best Regards,
--
Consultant
Dipl.-Inf. Uwe Grohnwaldt
Gutleutstr. 351
60327 Frankfurt a. M.

eMail: u...@grohnwaldt.eu
Telefon: +49-69-34878906
Mobil: +49-172-3209285
Fax: +49-69-348789069

- Original Message -
> From: "Andrei Mikhailovsky" 
> To: "Uwe Grohnwaldt" 
> Cc: ceph-users@lists.ceph.com
> Sent: Montag, 12. Mai 2014 14:48:58
> Subject: Re: [ceph-users] Ceph with VMWare / XenServer
> 
> 
> Uwe, thanks for your quick reply.
> 
> Do you run the Xenserver setup on production env and have you tried
> to test some failover scenarios to see if the xenserver guest vms
> are working during the failover of storage servers?
> 
> Also, how did you set up the xenserver iscsi? Have you used the
> multipath option to set up the LUNs?
> 
> Cheers
> 
> 
> 
> 
> - Original Message -
> 
> From: "Uwe Grohnwaldt" 
> To: ceph-users@lists.ceph.com
> Sent: Monday, 12 May, 2014 12:57:48 PM
> Subject: Re: [ceph-users] Ceph with VMWare / XenServer
> 
> Hi,
> 
> at the moment we are using tgt with RBD backend compiled from source
> on Ubuntu 12.04 and 14.04 LTS. We have two machines within two
> ip-ranges (e.g. 192.168.1.0/24 and 192.168.2.0/24). One machine in
> 192.168.1.0/24 and one machine in 192.168.2.0/24. The config for tgt
> is the same on both machines, they export the same rbd. This works
> well for XenServer.
> 
> For VMWare you have to disable VAAI to use it with tgt
> (http://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1033665)
> If you don't disable it, ESXi becomes very slow and unresponsive.
> 
> I think the problem is the iSCSI Write Same Support but I haven't
> tried which of the settings of VAAI is responsible for this
> behavior.
> 
> Mit freundlichen Grüßen / Best Regards,
> --
> Consultant
> Dipl.-Inf. Uwe Grohnwaldt
> Gutleutstr. 351
> 60327 Frankfurt a. M.
> 
> eMail: u...@grohnwaldt.eu
> Telefon: +49-69-34878906
> Mobil: +49-172-3209285
> Fax: +49-69-348789069
> 
> - Original Message -
> > From: "Andrei Mikhailovsky" 
> > To: ceph-users@lists.ceph.com
> > Sent: Montag, 12. Mai 2014 12:00:48
> > Subject: [ceph-users] Ceph with VMWare / XenServer
> > 
> > 
> > 
> > Hello guys,
> > 
> > I am currently running a ceph cluster for running vms with qemu +
> > rbd. It works pretty well and provides a good degree of failover. I
> > am able to run maintenance tasks on the ceph nodes without
> > interrupting vms IO.
> > 
> > I would like to do the same with VMWare / XenServer hypervisors,
> > but
> > I am not really sure how to achieve this. Initially I thought of
> > using iscsi multipathing, however, as it turns out, multipathing is
> > more for load balancing and nic/switch failure. It does not allow
> > me
> > to perform maintenance on the iscsi target without interrupting
> > service to vms.
> > 
> > Has anyone done either a PoC or better a production environment
> > where
> > they've used ceph as a backend storage with vmware / xenserver? The
> > important element for me is to have the ability of performing
> > maintenance tasks and resilience to failovers without interrupting
> > IO to vms. Are there any recommendations or howtos on how this
> > could
> > be achieved?
> > 
> > Many thanks
> > 
> > Andrei
> > 
> > 
> > ___
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > 
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 
> 
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Ceph with VMWare / XenServer

2014-05-12 Thread Jake Young
Hello Andrei,

I'm trying to accomplish the same thing with VMWare. So far I'm still doing
lab testing, but we've gotten as far as simulating a production workload.
 Forgive the lengthy reply, I happen to be sitting on an airplane .

My existing solution is using NFS servers running in ESXi VMs. Each VM
serves one or two large (2-4 TB) rbd images. These images are for vmdk
storage as well as oracle RAC disks.

I tested using multiple NFS servers serving a single rbd, but kept on
seeing xfs corruption (which was recoverable with xfs_repair). I initially
blamed ceph, but eventually realized that the problem is actually with xfs;
well in fact, the problem was with my configuration. It is generally a very
bad idea to write to the same xfs file system from two separate computers,
whether it is to a ceph rbd or to a physical disk in a shared disk array.
What would be required would be a way to synchronize writes between the
servers mounting the rbd. There are protocols available to do this, but all
of them would introduce more latency, which I'm already struggling to
control.

My environment is all Cisco UCS hardware. C240 rack mount servers for OSDs
and B200 blade servers for VMWare ESXi. The entire network is 10Gb or
better.  After carefully examining my nfs servers (which are VMs running in
ESXi on local storage), I found that I had a tremendous amount of kernel
IO. This was because of the high volume of TCP packets it had to constantly
process for BOTH the NFS traffic and the ceph traffic.

One thing that helped was to enable jumbo frames on every device in the
path from ESXi to the OSDs. This is not as simple as it sounds. In ESXi,
the vmk port and the vSwitch the vmk is on must have the mtu set to 9000.
In the switches, the VLANs and the interfaces need to have the mtu set to
9128 (don't forget about vlan tagging overhead). In the UCSM (Cisco GUI for
configuring the Blades and networking), all the vnics and the qos policies
must be set to 9000. The Linux interfaces in the nfs servers, mons, and
osds all needed to be set to 9000 as well.

My kernel io was still high, so I just gave the NFS VM more vCPUs (8
vCPUs, 8 GB RAM).  This helped as well.

With that all in place, my lab environment is doing a sustained 200 iops
bursting up to 500 iops (from VMWare's perspective) on one NFS server VM.
The IO is mostly small writes. My lab cluster just has 11 osds in a single
node.  I have 3x replication as well, so the cluster is actually doing more
like 600 - 1400 iops. The osds have an LSI 2208 controller (2GB cache) with
each disk in separate single disk RAID1 virtual drives (necessary to take
advantage of the write back cache). The OSDs have no separate journal;
which means the disks are actually writing at 1200 - 2800 iops (journal +
data). Not bad for one node with 11x 7k disks.

I still have high latency (though it is much better than before enabling
jumbo frames). VMWare shows between 10,000 microseconds and 200,000
microseconds of latency.  That is acceptable for this application.  IO is
mostly asynchronous: alarming/logging writes, database updates. I don't
notice the latency on the VMs running in the ceph-NFS datastore.

I believe the latency is actually from the osd node being pretty much maxed
out. I have 4 more osd servers on order to hopefully smooth out the latency
spikes.


One huge problem with the NFS server gateway approach is that you have many
layers of file systems that are introduced in each OS. My current
solution's file system stack looks like this:

ext4 - VMs file systems
VMFS - ESXi
NFS - between ESXi and nfs server
XFS - NFS server to mounted rbd disk
Rados - NFS server ceph kernel client to OSDs
XFS - OSDs to local file system

Yuck!  Four journaling file systems to write through: VMFS, XFS, OSD, XFS.


Clearly the best approach would be for the VMs to directly access the ceph
cluster:

ext4 - VMs file systems
Rados - VM ceph kernel client to OSDs
XFS - OSDs to local file system

Due to the packaging/deployment procedure of my application (and the
ancient RHEL 5 kernel), that won't be possible any time soon. The
application will be migrated to openstack, off of VMWare, first.

Since I'm using UCS hardware, there is native FCoE built in (with FC frame
offload and I can even boot off of FCoE); I am going to build a pair
of fiber channel gateways to replace the NFS server. The the filesystem
stack will look like this:

ext4 - VMs file systems
VMFS - ESXi
FC - between UCS vHBA and FC Target
Rados - FC target via LIO, ceph kernel client to OSDs
XFS - OSDs to local file system

I had some issues with getting a B200 blade to work in FC target mode (it
was only designed to be an initiator), so I'll have to use a C240 in
independent mode connected to a nexus 5k switch.

As an alternative (while I wait for my new osd nodes and nexus switches to
arrive), I was interested in trying tgt with fcoe. I've seen some negative
performance reports due to using userland ceph client vs kernel client.
More imp

Re: [ceph-users] Ceph with VMWare / XenServer

2014-05-12 Thread Andrei Mikhailovsky
Uwe, thanks for your quick reply. 

Do you run the Xenserver setup on production env and have you tried to test 
some failover scenarios to see if the xenserver guest vms are working during 
the failover of storage servers? 

Also, how did you set up the xenserver iscsi? Have you used the multipath 
option to set up the LUNs? 

Cheers 




- Original Message -

From: "Uwe Grohnwaldt"  
To: ceph-users@lists.ceph.com 
Sent: Monday, 12 May, 2014 12:57:48 PM 
Subject: Re: [ceph-users] Ceph with VMWare / XenServer 

Hi, 

at the moment we are using tgt with RBD backend compiled from source on Ubuntu 
12.04 and 14.04 LTS. We have two machines within two ip-ranges (e.g. 
192.168.1.0/24 and 192.168.2.0/24). One machine in 192.168.1.0/24 and one 
machine in 192.168.2.0/24. The config for tgt is the same on both machines, 
they export the same rbd. This works well for XenServer. 

For VMWare you have to disable VAAI to use it with tgt 
(http://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1033665)
 If you don't disable it, ESXi becomes very slow and unresponsive. 

I think the problem is the iSCSI Write Same Support but I haven't tried which 
of the settings of VAAI is responsible for this behavior. 

Mit freundlichen Grüßen / Best Regards, 
-- 
Consultant 
Dipl.-Inf. Uwe Grohnwaldt 
Gutleutstr. 351 
60327 Frankfurt a. M. 

eMail: u...@grohnwaldt.eu 
Telefon: +49-69-34878906 
Mobil: +49-172-3209285 
Fax: +49-69-348789069 

- Original Message - 
> From: "Andrei Mikhailovsky"  
> To: ceph-users@lists.ceph.com 
> Sent: Montag, 12. Mai 2014 12:00:48 
> Subject: [ceph-users] Ceph with VMWare / XenServer 
> 
> 
> 
> Hello guys, 
> 
> I am currently running a ceph cluster for running vms with qemu + 
> rbd. It works pretty well and provides a good degree of failover. I 
> am able to run maintenance tasks on the ceph nodes without 
> interrupting vms IO. 
> 
> I would like to do the same with VMWare / XenServer hypervisors, but 
> I am not really sure how to achieve this. Initially I thought of 
> using iscsi multipathing, however, as it turns out, multipathing is 
> more for load balancing and nic/switch failure. It does not allow me 
> to perform maintenance on the iscsi target without interrupting 
> service to vms. 
> 
> Has anyone done either a PoC or better a production environment where 
> they've used ceph as a backend storage with vmware / xenserver? The 
> important element for me is to have the ability of performing 
> maintenance tasks and resilience to failovers without interrupting 
> IO to vms. Are there any recommendations or howtos on how this could 
> be achieved? 
> 
> Many thanks 
> 
> Andrei 
> 
> 
> ___ 
> ceph-users mailing list 
> ceph-users@lists.ceph.com 
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
> 
___ 
ceph-users mailing list 
ceph-users@lists.ceph.com 
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Ceph with VMWare / XenServer

2014-05-12 Thread Uwe Grohnwaldt
Hi,

at the moment we are using tgt with RBD backend compiled from source on Ubuntu 
12.04 and 14.04 LTS. We have two machines within two ip-ranges (e.g. 
192.168.1.0/24 and 192.168.2.0/24). One machine in 192.168.1.0/24 and one 
machine in 192.168.2.0/24. The config for tgt is the same on both machines, 
they export the same rbd. This works well for XenServer.

For VMWare you have to disable VAAI to use it with tgt 
(http://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1033665)
 If you don't disable it, ESXi becomes very slow and unresponsive.

I think the problem is the iSCSI Write Same Support but I haven't tried which 
of the settings of VAAI is responsible for this behavior.

Mit freundlichen Grüßen / Best Regards,
--
Consultant
Dipl.-Inf. Uwe Grohnwaldt
Gutleutstr. 351
60327 Frankfurt a. M.

eMail: u...@grohnwaldt.eu
Telefon: +49-69-34878906
Mobil: +49-172-3209285
Fax: +49-69-348789069

- Original Message -
> From: "Andrei Mikhailovsky" 
> To: ceph-users@lists.ceph.com
> Sent: Montag, 12. Mai 2014 12:00:48
> Subject: [ceph-users] Ceph with VMWare / XenServer
> 
> 
> 
> Hello guys,
> 
> I am currently running a ceph cluster for running vms with qemu +
> rbd. It works pretty well and provides a good degree of failover. I
> am able to run maintenance tasks on the ceph nodes without
> interrupting vms IO.
> 
> I would like to do the same with VMWare / XenServer hypervisors, but
> I am not really sure how to achieve this. Initially I thought of
> using iscsi multipathing, however, as it turns out, multipathing is
> more for load balancing and nic/switch failure. It does not allow me
> to perform maintenance on the iscsi target without interrupting
> service to vms.
> 
> Has anyone done either a PoC or better a production environment where
> they've used ceph as a backend storage with vmware / xenserver? The
> important element for me is to have the ability of performing
> maintenance tasks and resilience to failovers without interrupting
> IO to vms. Are there any recommendations or howtos on how this could
> be achieved?
> 
> Many thanks
> 
> Andrei
> 
> 
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Ceph with VMWare / XenServer

2014-05-12 Thread Andrei Mikhailovsky
Hello guys, 

I am currently running a ceph cluster for running vms with qemu + rbd. It works 
pretty well and provides a good degree of failover. I am able to run 
maintenance tasks on the ceph nodes without interrupting vms IO. 

I would like to do the same with VMWare / XenServer hypervisors, but I am not 
really sure how to achieve this. Initially I thought of using iscsi 
multipathing, however, as it turns out, multipathing is more for load balancing 
and nic/switch failure. It does not allow me to perform maintenance on the 
iscsi target without interrupting service to vms. 

Has anyone done either a PoC or better a production environment where they've 
used ceph as a backend storage with vmware / xenserver? The important element 
for me is to have the ability of performing maintenance tasks and resilience to 
failovers without interrupting IO to vms. Are there any recommendations or 
howtos on how this could be achieved? 

Many thanks 

Andrei 

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com