Yiping,

Why not make a blog post about it so everyone can benefit? :)

Lucian

--
Sent from the Delta quadrant using Borg technology!

Nux!
www.nux.ro

----- Original Message -----
> From: "Yiping Zhang" <yzh...@marketo.com>
> To: users@cloudstack.apache.org, aemne...@gmail.com
> Sent: Friday, 8 January, 2016 01:31:21
> Subject: Re: A Story of a Failed XenServer Upgrade

> Hi, Alessandro
> 
> Late to the thread.  Is this still an issue for you ?
> 
> I went thru this process before and I have a step by step document that I can
> share if you still need it.
> 
> Yiping
> 
> 
> 
> 
> On 1/2/16, 4:43 PM, "Ahmad Emneina" <aemne...@gmail.com> wrote:
> 
>>Hi Alessandro,
>>Without seeing the logs, or DB, it will be hard to diagnose the issue. I've
>>seen something similar in the past, where the XenServer host version isnt
>>getting updated in the DB, as part of the XS upgrade process. That caused
>>CloudStack to use the wrong hypervisor resource to try connecting back to
>>the XenServers... ending up in failure. If you could share sanitized
>>versions of your log and db, someone here might be able to give you the
>>necessary steps to get your cluster back under CloudStack control.
>>
>>On Sat, Jan 2, 2016 at 1:27 PM, Alessandro Caviglione <
>>c.alessan...@gmail.com> wrote:
>>
>>> No guys,as the article wrote, my first action was to put in Maintenance
>>> Mode the Pool Master INSIDE CS; "It is vital that you upgrade the XenServer
>>> Pool Master first before any of the Slaves.  To do so you need to empty the
>>> Pool Master of all CloudStack VMs, and you do this by putting the Host into
>>> Maintenance Mode within CloudStack to trigger a live migration of all VMs
>>> to alternate Hosts"
>>>
>>> This is exactly what I've done and after the XS upgrade, no hosts was able
>>> to communicate with CS and also with the upgraded host.
>>>
>>> Putting an host in Maint Mode within CS will trigger MM also on XenServer
>>> host or just will move the VMs to other hosts?
>>>
>>> And again.... what's the best practices to upgrade a XS cluster?
>>>
>>> On Sat, Jan 2, 2016 at 7:11 PM, Remi Bergsma <rberg...@schubergphilis.com>
>>> wrote:
>>>
>>> > CloudStack should always do the migration of VM's not the Hypervisor.
>>> >
>>> > That's not true. You can safely migrate outside of CloudStack as the
>>> power
>>> > report will tell CloudStack where the vms live and the db gets updated
>>> > accordingly. I do this a lot while patching and that works fine on 6.2
>>> and
>>> > 6.5. I use both CloudStack 4.4.4 and 4.7.0.
>>> >
>>> > Regards, Remi
>>> >
>>> >
>>> > Sent from my iPhone
>>> >
>>> > On 02 Jan 2016, at 16:26, Jeremy Peterson <jpeter...@acentek.net<mailto:
>>> > jpeter...@acentek.net>> wrote:
>>> >
>>> > I don't use XenServer maintenance mode until after CloudStack has put the
>>> > Host in maintenance mode.
>>> >
>>> > When you initiate maintenance mode from the host rather than CloudStack
>>> > the db does not know where the VM's are and your UUID's get jacked.
>>> >
>>> > CS is your brains not the hypervisor.
>>> >
>>> > Maintenance in CS.  All VM's will migrate.  Maintenance in XenCenter.
>>> > Upgrade.  Reboot.  Join Pool.  Remove Maintenance starting at hypervisor
>>> if
>>> > needed and then CS and move on to the next Host.
>>> >
>>> > CloudStack should always do the migration of VM's not the Hypervisor.
>>> >
>>> > Jeremy
>>> >
>>> >
>>> > -----Original Message-----
>>> > From: Davide Pala [mailto:davide.p...@gesca.it]
>>> > Sent: Friday, January 1, 2016 5:18 PM
>>> > To: users@cloudstack.apache.org<mailto:users@cloudstack.apache.org>
>>> > Subject: R: A Story of a Failed XenServer Upgrade
>>> >
>>> > Hi alessandro. If u put in maintenance mode the master you force the
>>> > election of a new pool master. Now when you have see the upgraded host as
>>> > disconnected you are connected to the new pool master and the host (as a
>>> > pool member) cannot comunicate with a pool master of an earliest version.
>>> > The solution? Launche the upgrade on the pool master without enter in
>>> > maintenance mode. And remember a consistent backup!!!
>>> >
>>> >
>>> >
>>> > Inviato dal mio dispositivo Samsung
>>> >
>>> >
>>> > -------- Messaggio originale --------
>>> > Da: Alessandro Caviglione <c.alessan...@gmail.com<mailto:
>>> > c.alessan...@gmail.com>>
>>> > Data: 01/01/2016 23:23 (GMT+01:00)
>>> > A: users@cloudstack.apache.org<mailto:users@cloudstack.apache.org>
>>> > Oggetto: A Story of a Failed XenServer Upgrade
>>> >
>>> > Hi guys,
>>> > I want to share my XenServer Upgrade adventure to understand if I did
>>> > domething wrong.
>>> > I upgraded CS from 4.4.4 to 4.5.2 without any issues, after all the VRs
>>> > has been upgraded I start the upgrade process of my XenServer hosts from
>>> > 6.2 to 6.5.
>>> > I do not already have PoolHA enabled so I followed this article:
>>> >
>>> >
>>> http://www.shapeblue.com/how-to-upgrade-an-apache-cloudstack-citrix-xenserver-cluster/
>>> >
>>> > The cluster consists of n° 3 XenServer hosts.
>>> >
>>> > First of all I added manage.xenserver.pool.master=false
>>> > to environment.properties file and restarted cloudstack-management
>>> service.
>>> >
>>> > After that I put in Maintenance Mode Pool Master host and, after all VMs
>>> > has been migrated, I Unmanaged the cluster.
>>> > At this point all host appears as "Disconnected" from CS interface and
>>> > this should be right.
>>> > Now I put XenServer 6.5 CD in the host in Maintenance Mode and start a
>>> > in-place upgrade.
>>> > After XS6.5 has been installed, I istalled the 6.5SP1 and reboot again.
>>> > At this point I expected that, after click on Manage Cluster on CS, all
>>> > the hosts come back to "UP" and I could go ahead upgrading the other
>>> > hosts....
>>> >
>>> > But, instead of that, all the hosts still appears as "Disconnected", I
>>> > tried a couple of cloudstack-management service restart without success.
>>> >
>>> > So I opened XenCenter and connect to Pool Master I upgraded to 6.5 and it
>>> > appear in Maintenance Mode, so I tried to Exit from Maint Mode but I got
>>> > the error: The server is still booting
>>> >
>>> > After some investigation, I run the command "xe task-list" and this is
>>> the
>>> > result:
>>> >
>>> > uuid ( RO)                : 72f48a56-1d24-1ca3-aade-091f1830e2f1
>>> > name-label ( RO): VM.set_memory_dynamic_range name-description ( RO):
>>> > status ( RO): pending
>>> > progress ( RO): 0.000
>>> >
>>> > I tried a couple of reboot but nothing changes.... so I decided to shut
>>> > down the server, force raise a slave host to master with emergency mode,
>>> > remove old server from CS and reboot CS.
>>> >
>>> > After that, I see my cluster up and running again, so I installed XS
>>> > 6.2SP1 on the "upgraded" host and added again to the cluster....
>>> >
>>> > So after an entire day of work, I'm in the same situation! :D
>>> >
>>> > Anyone can tell me if I made something wrong??
>>> >
>>> > Thank you very much!
>>> >

Reply via email to