> Op 16 juli 2016 om 9:06 schreef Zoltan Arnold Nagy 
> <zol...@linux.vnet.ibm.com>:
> 
> 
> I’ve also upgraded last weekend our Hammer cluster to Jewel except for RGW. 
> Reading the initial user stories on the list where you had to run some 
> special script to create the default built-in zone discouraged me from doing 
> so.
> 
> So on .2, the rgw upgrade went without any hitch just by upgrading radosgw, 
> or did you run any helper scripts as well?
> 

I did not personally upgrade the RGW. Due to time constraints somebody else did 
that. In this case we had two RGWs in production we were upgraded.

When I asked the other day he said all went fine and no problems came up.

I checked afterwards and couldn't see a problem.

Wido

> Thanks!
> 
> > On 14 Jul 2016, at 08:33, Wido den Hollander <w...@42on.com> wrote:
> > 
> > To add, the RGWs upgraded just fine as well.
> > 
> > No regions in use here (yet!), so that upgraded as it should.
> > 
> > Wido
> > 
> >> Op 13 juli 2016 om 16:56 schreef Wido den Hollander <w...@42on.com>:
> >> 
> >> 
> >> Hello,
> >> 
> >> The last 3 days I worked at a customer with a 1800 OSD cluster which had 
> >> to be upgraded from Hammer 0.94.5 to Jewel 10.2.2
> >> 
> >> The cluster in this case is 99% RGW, but also some RBD.
> >> 
> >> I wanted to share some of the things we encountered during this upgrade.
> >> 
> >> All 180 nodes are running CentOS 7.1 on a IPv6-only network.
> >> 
> >> ** Hammer Upgrade **
> >> At first we upgraded from 0.94.5 to 0.94.7, this went well except for the 
> >> fact that the monitors got spammed with these kind of messages:
> >> 
> >> "Failed to encode map eXXX with expected crc"
> >> 
> >> Some searching on the list brought me to:
> >> 
> >> ceph tell osd.* injectargs -- --clog_to_monitors=false
> >> 
> >> This reduced the load on the 5 monitors and made recovery succeed smoothly.
> >> 
> >> ** Monitors to Jewel **
> >> The next step was to upgrade the monitors from Hammer to Jewel.
> >> 
> >> Using Salt we upgraded the packages and afterwards it was simple:
> >> 
> >>  killall ceph-mon
> >>  chown -R ceph:ceph /var/lib/ceph
> >>  chown -R ceph:ceph /var/log/ceph
> >> 
> >> Now, a systemd quirck. 'systemctl start ceph.target' does not work, I had 
> >> to manually enabled the monitor and start it:
> >> 
> >> systemctl enable ceph-mon@srv-zmb04-05.service
> >> systemctl start ceph-mon@srv-zmb04-05.service
> >> 
> >> Afterwards the monitors were running just fine.
> >> 
> >> ** OSDs to Jewel **
> >> To upgrade the OSDs to Jewel we initially used Salt to update the packages 
> >> on all systems to 10.2.2, we then used a Shell script which we ran on one 
> >> node at a time.
> >> 
> >> The failure domain here is 'rack', so we executed this in one rack, then 
> >> the next one, etc, etc.
> >> 
> >> Script can be found on Github: 
> >> https://gist.github.com/wido/06eac901bd42f01ca2f4f1a1d76c49a6
> >> 
> >> Be aware that the chown can take a long, long, very long time!
> >> 
> >> We ran into the issue that some OSDs crashed after start. But after trying 
> >> again they would start.
> >> 
> >> "void FileStore::init_temp_collections()"
> >> 
> >> I reported this in the tracker as I'm not sure what is happening here: 
> >> http://tracker.ceph.com/issues/16672
> >> 
> >> ** New OSDs with Jewel **
> >> We also had some new nodes which we wanted to add to the Jewel cluster.
> >> 
> >> Using Salt and ceph-disk we ran into a partprobe issue in combination with 
> >> ceph-disk. There was already a Pull Request for the fix, but that was not 
> >> included in Jewel 10.2.2.
> >> 
> >> We manually applied the PR and it fixed our issues: 
> >> https://github.com/ceph/ceph/pull/9330
> >> 
> >> Hope this helps other people with their upgrades to Jewel!
> >> 
> >> Wido
> >> _______________________________________________
> >> ceph-users mailing list
> >> ceph-users@lists.ceph.com
> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > 
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to