Hi Victor,

If I may interject, I read your email and understand you're running KVM
with Ceph storage. As I far I know, ACS only supports HA on NFS or iSCSI
primary storage.

http://docs.cloudstack.apache.org/projects/cloudstack-administration/en/4.11/reliability.html

However, if you wanted to use Ceph, you could create an RBD block device
and export it over NFS. Here is an article I referenced in the past:

https://www.sebastien-han.fr/blog/2012/07/06/nfs-over-rbd/

You could then add that NFS storage into ACS and utilize HA. I hope I'm
understanding you correctly.

Best Regards,
James

On Thu, Apr 5, 2018 at 12:53 PM, victor <vic...@ihnetworks.com> wrote:

> Hello Boris,
>
> I am able to create VM with nfs+Ha and nfs without HA. The issue is with
> creating VM with Ceph  storage.
>
> Regards
> Victor
>
>
>
> On 04/05/2018 01:18 PM, Boris Stoyanov wrote:
>
>> Hi Victor,
>> Host HA is working only with KVM + NFS. Ceph is not supported at this
>> stage. Obviously RAW volumes are not supported on your pool, but I’m not
>> sure if that’s because of Ceph or HA in general. Are you able to deploy a
>> non-ha VM?
>>
>> Boris Stoyanov
>>
>>
>> boris.stoya...@shapeblue.com
>> www.shapeblue.com
>> 53 Chandos Place, Covent Garden, London  WC2N 4HSUK
>> @shapeblue
>>
>>
>>> On 5 Apr 2018, at 4:19, victor <vic...@ihnetworks.com> wrote:
>>>
>>> Hello Rohit,
>>>
>>> Is the Host HA provider start working with Ceph. The reason I am asking
>>> is because, I am not able to create a VM with Ceph storage in a kvm host
>>> with HA enabled and I am getting the following error while creating VM.
>>>
>>> ============
>>> .cloud.exception.StorageUnavailableException: Resource [StoragePool:2]
>>> is unreachable: Unable to create Vol[9|vm=6|DATADISK]:com.cloud
>>> .utils.exception.CloudRuntimeException: org.libvirt.LibvirtException:
>>> unsupported configuration: only RAW volumes are supported by this storage
>>> pool
>>> ============
>>>
>>> Regards
>>> Victor
>>>
>>> On 11/04/2017 09:53 PM, Rohit Yadav wrote:
>>>
>>>> Hi James, (/cc Simon and others),
>>>>
>>>>
>>>> A new feature exists in upcoming ACS 4.11, Host HA:
>>>>
>>>> https://cwiki.apache.org/confluence/display/CLOUDSTACK/Host+HA
>>>>
>>>> You can read more about it here as well: http://www.shapeblue.com/host-
>>>> ha-for-kvm-hosts-in-cloudstack/
>>>>
>>>> This feature can use a custom HA provider, with default HA provider
>>>> implemented for KVM and NFS, and uses ipmi based fencing (STONITH) of the
>>>> host. The current HA mechanism provides no such method of fencing (powering
>>>> off) a host and it depends under what circumstances the VM HA is failing
>>>> (environment issues, ACS version etc).
>>>>
>>>> As Simon mentioned, we have a (host) HA provider that works with Ceph
>>>> in near future.
>>>>
>>>> Regards.
>>>>
>>>> ________________________________
>>>> From: Simon Weller <swel...@ena.com.INVALID>
>>>> Sent: Thursday, November 2, 2017 7:27:22 PM
>>>> To: users@cloudstack.apache.org
>>>> Subject: Re: Problems with KVM HA & STONITH
>>>>
>>>> James,
>>>>
>>>>
>>>> Ceph is a great solution and we run all of our ACS storage on Ceph.
>>>> Note that it adds another layer of complexity to your installation, so
>>>> you're going need to develop some expertise with that platform to get
>>>> comfortable with how it works. Typically you don't want to mix Ceph with
>>>> your ACS hosts. We in fact deploy 3 separate Ceph Monitors, and then scale
>>>> OSDs as required on a per cluster basis in order to add additional
>>>> resiliency (So every KVM ACS cluster has it's own Ceph "POD").  We also use
>>>> Ceph for S3 storage (on completely separate Ceph clusters) for some other
>>>> services.
>>>>
>>>>
>>>> NFS is much simpler to maintain for smaller installations in my
>>>> opinion. If the IO load you're looking at isn't going to be insanely high,
>>>> you could look at building a 2 node NFS cluster using pacemaker and DRDB
>>>> for data replication between nodes. That would reduce your storage
>>>> requirement to 2 fairly low power servers (NFS is not very cpu intensive).
>>>> Currently on a host failure when using a storage other than NFS on KVM, you
>>>> will not see HA occur until you take the failed host out of the ACS
>>>> cluster. This is a historical limitation because ACS could not confirm the
>>>> host had been fenced correctly, so to avoid potential data corruption (due
>>>> to 2 hosts mounting the same storage), it doesn't do anything until the
>>>> operator intervenes. As of ACS 4.10, IPMI based fencing is now supported on
>>>> NFS and we're planning on developing similar support for Ceph.
>>>>
>>>>
>>>> Since you're an school district, I'm more than happy to jump on the
>>>> phone with you to talk you through these options if you'd like.
>>>>
>>>>
>>>> - Si
>>>>
>>>>
>>>> ________________________________
>>>> From: McClune, James <mcclu...@norwalktruckers.net>
>>>> Sent: Thursday, November 2, 2017 8:28 AM
>>>> To: users@cloudstack.apache.org
>>>> Subject: Re: Problems with KVM HA & STONITH
>>>>
>>>> Hi Simon,
>>>>
>>>> Thanks for getting back to me. I created one single NFS share and added
>>>> it
>>>> as primary storage. I think I better understand how the storage works,
>>>> with
>>>> ACS.
>>>>
>>>> I was able to get HA working with one NFS storage, which is good.
>>>> However,
>>>> is there a way to incorporate multiple NFS storage pools and still have
>>>> the
>>>> HA functionality? I think something like GlusterFS or Ceph (like Ivan
>>>> and
>>>> Dag described) will work better.
>>>>
>>>> Thank you Simon, Ivan, and Dag for your assistance!
>>>> James
>>>>
>>>> On Wed, Nov 1, 2017 at 10:10 AM, Simon Weller <swel...@ena.com.invalid>
>>>> wrote:
>>>>
>>>> James,
>>>>>
>>>>>
>>>>> Try just configuring a single NFS server and see if your setup works.
>>>>> If
>>>>> you have 3 NFS shares, across all 3 hosts, i'm wondering whether ACS is
>>>>> picking the one you rebooted as the storage for your VMs and when that
>>>>> storage goes away (when you bounce the host), all storage for your VMs
>>>>> vanishes and ACS tries to reboot your other hosts.
>>>>>
>>>>>
>>>>> Normally in a simple ACS setup, you would have a separate storage
>>>>> server
>>>>> that can serve up NFS to all hosts. If a host dies, then a VM would be
>>>>> brought up on a spare hosts since all hosts have access to the same
>>>>> storage.
>>>>>
>>>>> Your other option is to use local storage, but that won't provide HA.
>>>>>
>>>>>
>>>>> - Si
>>>>>
>>>>>
>>>>> ________________________________
>>>>> From: McClune, James <mcclu...@norwalktruckers.net>
>>>>> Sent: Monday, October 30, 2017 2:26 PM
>>>>> To: users@cloudstack.apache.org
>>>>> Subject: Re: Problems with KVM HA & STONITH
>>>>>
>>>>> Hi Dag,
>>>>>
>>>>> Thank you for responding back. I am currently running ACS 4.9 on an
>>>>> Ubuntu
>>>>> 14.04 VM. I have the three nodes, each having about 1TB of primary
>>>>> storage
>>>>> (NFS) and 1TB of secondary storage (NFS). I added each NFS share into
>>>>> ACS.
>>>>> All nodes are in a cluster.
>>>>>
>>>>> Maybe I'm not understanding the setup or misconfigured something. I'm
>>>>> trying to setup an HA environment where if one node goes down, running
>>>>> an
>>>>> HA marked VM, the VM will start on another host. When I simulate a
>>>>> network
>>>>> disconnect or reboot of a host, all of the nodes go down (STONITH?).
>>>>>
>>>>> I am unsure on how to setup an HA environment, if all the nodes in the
>>>>> cluster go down. Any help is much appreciated!
>>>>>
>>>>> Thanks,
>>>>> James
>>>>>
>>>>> On Mon, Oct 30, 2017 at 3:49 AM, Dag Sonstebo <
>>>>> dag.sonst...@shapeblue.com>
>>>>> wrote:
>>>>>
>>>>> Hi James,
>>>>>>
>>>>>> I think  you possibly have over-configured your KVM hosts. If you use
>>>>>> NFS
>>>>>> (and no clustered file system like CLVM) then there should be no need
>>>>>> to
>>>>>> configure STONITH. CloudStack takes care of your HA, so this is not
>>>>>> something you offload to the KVM host.
>>>>>>
>>>>>> (As mentioned the only time I have played with STONITH and CloudStack
>>>>>> was
>>>>>> for CLVM – and I eventually found it not fit for purpose, too unstable
>>>>>>
>>>>> and
>>>>>
>>>>>> causing too many issues like you describe. Note this was for block
>>>>>>
>>>>> storage
>>>>>
>>>>>> though – not NFS).
>>>>>>
>>>>>> Regards,
>>>>>> Dag Sonstebo
>>>>>> Cloud Architect
>>>>>> ShapeBlue
>>>>>>
>>>>>> On 28/10/2017, 03:40, "Ivan Kudryavtsev" <kudryavtsev...@bw-sw.com>
>>>>>>
>>>>> wrote:
>>>>>
>>>>>>      Hi. If the node losts nfs host it reboots (acs agent behaviour).
>>>>>> If
>>>>>>
>>>>> you
>>>>>
>>>>>>      really have 3 storages, you'll go clusterwide reboot everytime
>>>>>> your
>>>>>> host is
>>>>>>      down.
>>>>>>
>>>>>>      28 окт. 2017 г. 3:02 пользователь "Simon Weller"
>>>>>> <swel...@ena.com.invalid>
>>>>>>      написал:
>>>>>>
>>>>>>      > Hi James,
>>>>>>      >
>>>>>>      >
>>>>>>      > Can you elaborate a bit further on the storage? You say you're
>>>>>> running NFS
>>>>>>      > on all 3 nodes, can you explain how it is setup?
>>>>>>      >
>>>>>>      > Also, what version of ACS are you running?
>>>>>>      >
>>>>>>      >
>>>>>>      > - Si
>>>>>>      >
>>>>>>      >
>>>>>>      >
>>>>>>      >
>>>>>>      > ________________________________
>>>>>>      > From: McClune, James <mcclu...@norwalktruckers.net>
>>>>>>      > Sent: Friday, October 27, 2017 2:21 PM
>>>>>>      > To: users@cloudstack.apache.org
>>>>>>      > Subject: Problems with KVM HA & STONITH
>>>>>>      >
>>>>>>      > Hello Apache CloudStack Community,
>>>>>>      >
>>>>>>      > My setup consists of the following:
>>>>>>      >
>>>>>>      > - Three nodes (NODE1, NODE2, and NODE3)
>>>>>>      > NODE1 is running Ubuntu 16.04.3, NODE2 is running Ubuntu
>>>>>> 16.04.3,
>>>>>> and NODE3
>>>>>>      > is running Ubuntu 14.04.5.
>>>>>>      > - Management Server (running on separate VM, not in cluster)
>>>>>>      >
>>>>>>      > The three nodes use KVM as the hypervisor. I also configured
>>>>>>
>>>>> primary
>>>>>
>>>>>> and
>>>>>>      > secondary storage on all three of the nodes. I'm using NFS for
>>>>>> the
>>>>>> primary
>>>>>>      > & secondary storage. VM operations work great. Live migration
>>>>>> works
>>>>>> great.
>>>>>>      >
>>>>>>      > However, when a host goes down, the HA functionality does not
>>>>>> work
>>>>>> at all.
>>>>>>      > Instead of spinning up the VM on another available host, the
>>>>>> down
>>>>>> host
>>>>>>      > seems to trigger STONITH. When STONITH happens, all hosts in
>>>>>> the
>>>>>> cluster go
>>>>>>      > down. This not only causes no HA, but also downs perfectly good
>>>>>> VM's. I
>>>>>>      > have read countless articles and documentation related to this
>>>>>> issue. I
>>>>>>      > still cannot find a viable solution for this issue. I really
>>>>>> want
>>>>>>
>>>>> to
>>>>>
>>>>>> use
>>>>>>      > Apache CloudStack, but cannot implement this in production when
>>>>>> STONITH
>>>>>>      > happens.
>>>>>>      >
>>>>>>      > I think I have something misconfigured. I thought I would
>>>>>> reach out
>>>>>> to the
>>>>>>      > CloudStack community and ask for some friendly assistance.
>>>>>>      >
>>>>>>      > If there is anything (system-wise) you request in order to
>>>>>> further
>>>>>>      > troubleshoot this issue, please let me know and I'll send. I
>>>>>> appreciate any
>>>>>>      > help in this issue!
>>>>>>      >
>>>>>>      > --
>>>>>>      >
>>>>>>      > Thanks,
>>>>>>      >
>>>>>>      > James
>>>>>>      >
>>>>>>
>>>>>>
>>>>>>
>>>>>> dag.sonst...@shapeblue.com
>>>>>> www.shapeblue.com<http://www.shapeblue.com>
>>>>>>
>>>>> [http://www.shapeblue.com/wp-content/uploads/2017/06/logo.png]<
>>>> http://www.shapeblue.com/>
>>>>
>>>> Shapeblue - The CloudStack Company<http://www.shapeblue.com/>
>>>> www.shapeblue.com<http://www.shapeblue.com>
>>>> Rapid deployment framework for Apache CloudStack IaaS Clouds. CSForge
>>>> is a framework developed by ShapeBlue to deliver the rapid deployment of a
>>>> standardised ...
>>>>
>>>>
>>>>
>>>> [http://www.shapeblue.com/wp-content/uploads/2017/06/logo.png
>>>>>
>>>> [http://www.shapeblue.com/wp-content/uploads/2017/06/logo.png]
>>>>
>>>> ]<
>>>>
>>>>> http://www.shapeblue.com/>
>>>>>
>>>> [http://www.shapeblue.com/wp-content/uploads/2017/06/logo.png]<
>>>> http://www.shapeblue.com/>
>>>>
>>>> Shapeblue - The CloudStack Company
>>>> <https://maps.google.com/?q=lue+-+The+CloudStack+Company&entry=gmail&source=g>
>>>> <http://www.shapeblue.com/>
>>>> www.shapeblue.com<http://www.shapeblue.com>
>>>> Rapid deployment framework for Apache CloudStack IaaS Clouds. CSForge
>>>> is a framework developed by ShapeBlue to deliver the rapid deployment of a
>>>> standardised ...
>>>>
>>>>
>>>>
>>>> Shapeblue - The CloudStack Company<http://www.shapeblue.com/>
>>>>>
>>>> [http://www.shapeblue.com/wp-content/uploads/2017/06/logo.png]<
>>>> http://www.shapeblue.com/>
>>>>
>>>> Shapeblue - The CloudStack Company
>>>> <https://maps.google.com/?q=CloudStack+Company&entry=gmail&source=g><
>>>> http://www.shapeblue.com/>
>>>> www.shapeblue.com<http://www.shapeblue.com>
>>>> Rapid deployment framework for Apache CloudStack IaaS Clouds. CSForge
>>>> is a framework developed by ShapeBlue to deliver the rapid deployment of a
>>>> standardised ...
>>>>
>>>>
>>>>
>>>> www.shapeblue.com<http://www.shapeblue.com>
>>>>>
>>>> [http://www.shapeblue.com/wp-content/uploads/2017/06/logo.png]<
>>>> http://www.shapeblue.com/>
>>>>
>>>> Shapeblue - The CloudStack Company<http://www.shapeblue.com/>
>>>> www.shapeblue.com<http://www.shapeblue.com>
>>>> Rapid deployment framework for Apache CloudStack IaaS Clouds. CSForge
>>>> is a framework developed by ShapeBlue to deliver the rapid deployment of a
>>>> standardised ...
>>>>
>>>>
>>>>
>>>> Rapid deployment framework for Apache CloudStack IaaS Clouds. CSForge
>>>>> is a
>>>>> framework developed by ShapeBlue to deliver the rapid deployment of a
>>>>> standardised ...
>>>>>
>>>>>
>>>>>
>>>>> 53 Chandos Place, Covent Garden, London  WC2N 4HSUK
>>>>>> @shapeblue
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>
>>>>>
>>>>>
>>>>> James McClune
>>>>>
>>>>> Technical Support Specialist
>>>>>
>>>>> Norwalk City Schools
>>>>>
>>>>> Phone: 419-660-6590
>>>>>
>>>>> mcclu...@norwalktruckers.net
>>>>>
>>>>>
>>>> --
>>>>
>>>>
>>>>
>>>> James McClune
>>>>
>>>> Technical Support Specialist
>>>>
>>>> Norwalk City Schools
>>>>
>>>> Phone: 419-660-6590
>>>>
>>>> mcclu...@norwalktruckers.net
>>>>
>>>> rohit.ya...@shapeblue.com
>>>> www.shapeblue.com
>>>> 53 Chandos Place, Covent Garden, London  WC2N 4HSUK
>>>> @shapeblue
>>>>
>>>>
>>>
>


-- 



James McClune

Technical Support Specialist

Norwalk City Schools

Phone: 419-660-6590

mcclu...@norwalktruckers.net

Reply via email to