I tend to not do too much each time: either upgrade or data migrate. The actual upgrade process is seamless... So you can just as easily upgrade the "current" cluster to hammer, and add/remove nodes on the fly. All of this is quite seamless and straightforward (other than the data migration itself).
On Wed, Aug 26, 2015 at 3:17 PM, Chang, Fangzhe (Fangzhe) < fangzhe.ch...@alcatel-lucent.com> wrote: > Thanks, Luis. > > > > The motivation for using the newer version is to keep up-to-date with Ceph > development, since we suspect the old versioned radosgw could not be > restarted possibly due to library mismatch. > > Do you know whether the self-healing feature of ceph is applicable between > different versions or not? > > > > Fangzhe > > > > *From:* Luis Periquito [mailto:periqu...@gmail.com] > *Sent:* Wednesday, August 26, 2015 10:11 AM > *To:* Chang, Fangzhe (Fangzhe) > *Cc:* ceph-users@lists.ceph.com > *Subject:* Re: [ceph-users] Migrating data into a newer ceph instance > > > > I Would say the easiest way would be to leverage all the self-healing of > ceph: add the new nodes to the old cluster, allow or force all the data to > migrate between nodes, and then remove the old ones out. > > > > Well to be fair you could probably just install radosgw on another node > and use it as your gateway without the need to even create a new OSD node... > > > > Or was there a reason to create a new cluster? I can tell you that one of > the clusters I have has been around since bobtail, and now it's hammer... > > > > On Wed, Aug 26, 2015 at 2:50 PM, Chang, Fangzhe (Fangzhe) < > fangzhe.ch...@alcatel-lucent.com> wrote: > > Hi, > > > > We have been running Ceph/Radosgw version 0.80.7 (Giant) and stored quite > some amount of data in it. We are only using ceph as an object store via > radosgw. Last week cheph-radosgw daemon suddenly refused to start (with > logs only showing “initialization timeout” error on Centos 7). This > triggers me to install a newer instance --- Ceph/Radosgw version 0.94.2 > (Hammer). The new instance has a different set of key rings by default. The > next step is to have all the data migrated. Does anyone know how to get the > existing data out of the old ceph cluster (Giant) and into the new > instance (Hammer)? Please note that in the old three-node cluster ceph osd > is still running but radosgw is not. Any suggestion will be greatly > appreciated. > > Thanks. > > > > Regards, > > > > Fangzhe Chang > > > > > > > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com