destroy did remove the auth key, however create didnt add the auth, I had to do it manually. Then I tried to start the osd.0 again and it failed because osdmap said it was destroyed.
I've summed my steps below: Here are my commands prior to create: root@int1:~# ceph osd out 0 <-- wait for rebalance/recover --> root@int1:~# ceph osd safe-to-destroy 0 OSD(s) 0 are safe to destroy without reducing data durability. root@int1:~# systemctl kill ceph-osd@0 root@int1:~# ceph status cluster: id: efad7df8-721d-43d8-8d02-449406e70b90 health: HEALTH_OK services: mon: 3 daemons, quorum int1,int2,int3 mgr: int1(active), standbys: int3, int2 osd: 6 osds: 5 up, 5 in data: pools: 2 pools, 320 pgs objects: 97038 objects, 364 GB usage: 1096 GB used, 1128 GB / 2224 GB avail pgs: 320 active+clean io: client: 289 kB/s rd, 870 kB/s wr, 46 op/s rd, 48 op/s wr root@int1:~# mount | grep /var/lib/ceph/osd/ceph-0 /dev/sdc1 on /var/lib/ceph/osd/ceph-0 type xfs (rw,noatime,attr2,inode64,noquota) root@int1:~# umount /var/lib/ceph/osd/ceph-0 root@int1:~# ceph-volume lvm zap /dev/sdc Zapping: /dev/sdc Running command: sudo wipefs --all /dev/sdc stdout: /dev/sdc: 8 bytes were erased at offset 0x00000200 (gpt): 45 46 49 20 50 41 52 54 /dev/sdc: 8 bytes were erased at offset 0x6fc86d5e00 (gpt): 45 46 49 20 50 41 52 54 /dev/sdc: 2 bytes were erased at offset 0x000001fe (PMBR): 55 aa /dev/sdc: calling ioctl to re-read partition table: Success Running command: dd if=/dev/zero of=/dev/sdc bs=1M count=10 stderr: 10+0 records in 10+0 records out 10485760 bytes (10 MB) copied stderr: , 0.0253999 s, 413 MB/s --> Zapping successful for: /dev/sdc root@int1:~# ceph osd destroy 0 --yes-i-really-mean-it destroyed osd.0 root@int1:~# ceph status cluster: id: efad7df8-721d-43d8-8d02-449406e70b90 health: HEALTH_OK services: mon: 3 daemons, quorum int1,int2,int3 mgr: int1(active), standbys: int3, int2 osd: 6 osds: 5 up, 5 in data: pools: 2 pools, 320 pgs objects: 97038 objects, 364 GB usage: 1096 GB used, 1128 GB / 2224 GB avail pgs: 320 active+clean io: client: 56910 B/s rd, 1198 kB/s wr, 15 op/s rd, 48 op/s wr root@int1:~# ceph-volume create --bluestore --data /dev/sdc --osd-id 0 usage: ceph-volume [-h] [--cluster CLUSTER] [--log-level LOG_LEVEL] [--log-path LOG_PATH] ceph-volume: error: unrecognized arguments: create --bluestore --data /dev/sdc --osd-id 0 root@int1:~# ceph-volume lvm create --bluestore --data /dev/sdc --osd-id 0 Running command: sudo vgcreate --force --yes ceph-efad7df8-721d-43d8-8d02-449406e70b90 /dev/sdc stderr: WARNING: lvmetad is running but disabled. Restart lvmetad before enabling it! stdout: Physical volume "/dev/sdc" successfully created stdout: Volume group "ceph-efad7df8-721d-43d8-8d02-449406e70b90" successfully created Running command: sudo lvcreate --yes -l 100%FREE -n osd-block-138ce507-f28a-45bf-814c-7fa124a9d9b9 ceph-efad7df8-721d-43d8-8d02-449406e70b90 stderr: WARNING: lvmetad is running but disabled. Restart lvmetad before enabling it! stdout: Logical volume "osd-block-138ce507-f28a-45bf-814c-7fa124a9d9b9" created. Running command: sudo mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-0 Running command: chown -R ceph:ceph /dev/dm-4 Running command: sudo ln -s /dev/ceph-efad7df8-721d-43d8-8d02-449406e70b90/osd-block-138ce507-f28a-45bf-814c-7fa124a9d9b9 /var/lib/ceph/osd/ceph-0/block Running command: sudo ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-0/activate.monmap stderr: got monmap epoch 2 Running command: ceph-authtool /var/lib/ceph/osd/ceph-0/keyring --create-keyring --name osd.0 --add-key AQA5Qmta9LERFhAAKU+AmT1Sm56nk7sWx2BATQ== stdout: creating /var/lib/ceph/osd/ceph-0/keyring stdout: added entity osd.0 auth auth(auid = 18446744073709551615 key=AQA5Qmta9LERFhAAKU+AmT1Sm56nk7sWx2BATQ== with 0 caps) Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/keyring Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/ Running command: sudo ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 0 --monmap /var/lib/ceph/osd/ceph-0/activate.monmap --key **************************************** --osd-data /var/lib/ceph/osd/ceph-0/ --osd-uuid 138ce507-f28a-45bf-814c-7fa124a9d9b9 --setuser ceph --setgroup ceph stderr: 2018-01-26 14:59:10.039549 7fd7ef951cc0 -1 bluestore(/var/lib/ceph/osd/ceph-0//block) _read_bdev_label unable to decode label at offset 102: buffer::malformed_input: void bluestore_bdev_label_t::decode(ceph::buffer::list::iterator&) decode past end of struct encoding stderr: 2018-01-26 14:59:10.039744 7fd7ef951cc0 -1 bluestore(/var/lib/ceph/osd/ceph-0//block) _read_bdev_label unable to decode label at offset 102: buffer::malformed_input: void bluestore_bdev_label_t::decode(ceph::buffer::list::iterator&) decode past end of struct encoding stderr: 2018-01-26 14:59:10.039925 7fd7ef951cc0 -1 bluestore(/var/lib/ceph/osd/ceph-0//block) _read_bdev_label unable to decode label at offset 102: buffer::malformed_input: void bluestore_bdev_label_t::decode(ceph::buffer::list::iterator&) decode past end of struct encoding stderr: 2018-01-26 14:59:10.039984 7fd7ef951cc0 -1 bluestore(/var/lib/ceph/osd/ceph-0/) _read_fsid unparsable uuid stderr: 2018-01-26 14:59:11.359951 7fd7ef951cc0 -1 key AQA5Qmta9LERFhAAKU+AmT1Sm56nk7sWx2BATQ== stderr: 2018-01-26 14:59:11.888476 7fd7ef951cc0 -1 created object store /var/lib/ceph/osd/ceph-0/ for osd.0 fsid efad7df8-721d-43d8-8d02-449406e70b90 Running command: sudo ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-efad7df8-721d-43d8-8d02-449406e70b90/osd-block-138ce507-f28a-45bf-814c-7fa124a9d9b9 --path /var/lib/ceph/osd/ceph-0 Running command: sudo ln -snf /dev/ceph-efad7df8-721d-43d8-8d02-449406e70b90/osd-block-138ce507-f28a-45bf-814c-7fa124a9d9b9 /var/lib/ceph/osd/ceph-0/block Running command: chown -R ceph:ceph /dev/dm-4 Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-0 Running command: sudo systemctl enable ceph-volume@lvm-0-138ce507-f28a-45bf-814c-7fa124a9d9b9 stderr: Created symlink from /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-0-138ce507-f28a-45bf-814c-7fa124a9d9b9.service to /lib/systemd/system/ceph-volume@.service. Running command: sudo systemctl start ceph-osd@0 root@int1:~# ceph status cluster: id: efad7df8-721d-43d8-8d02-449406e70b90 health: HEALTH_OK services: mon: 3 daemons, quorum int1,int2,int3 mgr: int1(active), standbys: int3, int2 osd: 6 osds: 5 up, 5 in data: pools: 2 pools, 320 pgs objects: 97038 objects, 364 GB usage: 1095 GB used, 1128 GB / 2224 GB avail pgs: 320 active+clean io: client: 294 kB/s rd, 1827 kB/s wr, 61 op/s rd, 96 op/s wr root@int1:~# ceph osd tree ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF -1 2.60458 root default -2 0.86819 host int1 0 ssd 0.43159 osd.0 destroyed 0 1.00000 3 ssd 0.43660 osd.3 up 1.00000 1.00000 -3 0.86819 host int2 1 ssd 0.43159 osd.1 up 1.00000 1.00000 4 ssd 0.43660 osd.4 up 1.00000 1.00000 -4 0.86819 host int3 2 ssd 0.43159 osd.2 up 1.00000 1.00000 5 ssd 0.43660 osd.5 up 1.00000 1.00000 root@int1:~# ceph auth ls Does not list osd.0 root@int1:~# ceph auth add osd.0 osd 'allow *' mon 'allow profile osd' mgr 'allow profile osd' -i /var/lib/ceph/osd/ceph-0/keyring added key for osd.0 root@int1:~# systemctl start ceph-osd@0 root@int1:~# ceph status cluster: id: efad7df8-721d-43d8-8d02-449406e70b90 health: HEALTH_OK services: mon: 3 daemons, quorum int1,int2,int3 mgr: int1(active), standbys: int3, int2 osd: 6 osds: 5 up, 5 in data: pools: 2 pools, 320 pgs objects: 97163 objects, 365 GB usage: 1097 GB used, 1127 GB / 2224 GB avail pgs: 320 active+clean io: client: 284 kB/s rd, 539 kB/s wr, 32 op/s rd, 30 op/s wr root@int1:~# systemctl status ceph-osd@0 ● ceph-osd@0.service - Ceph object storage daemon osd.0 Loaded: loaded (/lib/systemd/system/ceph-osd@.service; enabled) Drop-In: /lib/systemd/system/ceph-osd@.service.d └─ceph-after-pve-cluster.conf Active: inactive (dead) since Fri 2018-01-26 17:02:08 UTC; 54s ago Process: 6857 ExecStart=/usr/bin/ceph-osd -f --cluster ${CLUSTER} --id %i --setuser ceph --setgroup ceph (code=exited, status=0/SUCCESS) Process: 6851 ExecStartPre=/usr/lib/ceph/ceph-osd-prestart.sh --cluster ${CLUSTER} --id %i (code=exited, status=0/SUCCESS) Main PID: 6857 (code=exited, status=0/SUCCESS) Jan 26 17:02:08 int1 systemd[1]: Started Ceph object storage daemon osd.0. Jan 26 17:02:08 int1 ceph-osd[6857]: starting osd.0 at - osd_data /var/lib/ceph/osd/ceph-0 /var/lib/ceph/osd/ceph-0/journal Jan 26 17:02:08 int1 ceph-osd[6857]: 2018-01-26 17:02:08.801761 7fed0b5bbcc0 -1 osd.0 0 log_to_monitors {default=true} Jan 26 17:02:08 int1 ceph-osd[6857]: 2018-01-26 17:02:08.804600 7fecf2ee4700 -1 osd.0 0 waiting for initial osdmap Jan 26 17:02:08 int1 ceph-osd[6857]: 2018-01-26 17:02:08.909237 7fecf7eee700 -1 osd.0 1040 osdmap says I am destroyed, exiting After this I followed Reed Dier's steps somewhat, zapped the disk again, removed auth, crush, osd. Zapped the disk/parts and device mapper. Could then run the create command without issues. Kind Regards, David Majchrzak > 26 jan. 2018 kl. 18:56 skrev Wido den Hollander <w...@42on.com>: > > > > On 01/26/2018 06:53 PM, David Majchrzak wrote: >> I did do that. >> It didn't add the auth key to ceph, so I had to do that manually. Then it >> said that osd.0 was set as destroyed, which yes, it was still in crushmap. >> I followed the docs to a point. > > Odd, the 'destroy' command should remove the auth key. Afterwards ceph-volume > will use the bootstrap-osd key to create it again. > > I didn't try this with ceph-volume yet, but I'm in the process of doing the > same with ceph-disk going to BlueStore and that works just fine. > > Wido > >>> 26 jan. 2018 kl. 18:50 skrev Wido den Hollander <w...@42on.com>: >>> >>> >>> >>> On 01/26/2018 06:37 PM, David Majchrzak wrote: >>>> Ran: >>>> ceph auth del osd.0 >>>> ceph auth del osd.6 >>>> ceph auth del osd.7 >>>> ceph osd rm osd.0 >>>> ceph osd rm osd.6 >>>> ceph osd rm osd.7 >>>> which seems to have removed them. >>> >>> Did you destroy the OSD prior to running ceph-volume? >>> >>> $ ceph osd destroy 6 >>> >>> After you've done that you can use ceph-volume to re-create the OSD. >>> >>> Wido >>> >>>> Thanks for the help Reed! >>>> Kind Regards, >>>> David Majchrzak >>>>> 26 jan. 2018 kl. 18:32 skrev David Majchrzak <da...@visions.se >>>>> <mailto:da...@visions.se>>: >>>>> >>>>> Thanks that helped! >>>>> >>>>> Since I had already "halfway" created a lvm volume I wanted to start from >>>>> the beginning and zap it. >>>>> >>>>> Tried to zap the raw device but failed since --destroy doesn't seem to be >>>>> in 12.2.2 >>>>> >>>>> http://docs.ceph.com/docs/master/ceph-volume/lvm/zap/ >>>>> >>>>> root@int1:~# ceph-volume lvm zap /dev/sdc --destroy >>>>> usage: ceph-volume lvm zap [-h] [DEVICE] >>>>> ceph-volume lvm zap: error: unrecognized arguments: --destroy >>>>> >>>>> So i zapped it with the vg/lvm instead. >>>>> ceph-volume lvm zap >>>>> /dev/ceph-efad7df8-721d-43d8-8d02-449406e70b90/osd-block-138ce507-f28a-45bf-814c-7fa124a9d9b9 >>>>> >>>>> However I run create on it since the LVM was already there. >>>>> So I zapped it with sgdisk and ran dmsetup remove. After that I was able >>>>> to create it again. >>>>> >>>>> However - each "ceph-volume lvm create" that I ran that failed, >>>>> successfully added an osd to crush map ;) >>>>> >>>>> So I've got this now: >>>>> >>>>> root@int1:~# ceph osd df tree >>>>> ID CLASS WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS TYPE NAME >>>>> -1 2.60959 - 2672G 1101G 1570G 41.24 1.00 - root default >>>>> -2 0.87320 - 894G 369G 524G 41.36 1.00 - host int1 >>>>> 3 ssd 0.43660 1.00000 447G 358G 90295M 80.27 1.95 301 osd.3 >>>>> 8 ssd 0.43660 1.00000 447G 11273M 436G 2.46 0.06 19 osd.8 >>>>> -3 0.86819 - 888G 366G 522G 41.26 1.00 - host int2 >>>>> 1 ssd 0.43159 1.00000 441G 167G 274G 37.95 0.92 147 osd.1 >>>>> 4 ssd 0.43660 1.00000 447G 199G 247G 44.54 1.08 173 osd.4 >>>>> -4 0.86819 - 888G 365G 523G 41.09 1.00 - host int3 >>>>> 2 ssd 0.43159 1.00000 441G 193G 248G 43.71 1.06 174 osd.2 >>>>> 5 ssd 0.43660 1.00000 447G 172G 274G 38.51 0.93 146 osd.5 >>>>> 0 0 0 0 0 0 0 0 0 osd.0 >>>>> 6 0 0 0 0 0 0 0 0 osd.6 >>>>> 7 0 0 0 0 0 0 0 0 osd.7 >>>>> >>>>> I guess I can just remove them from crush,auth and rm them? >>>>> >>>>> Kind Regards, >>>>> >>>>> David Majchrzak >>>>> >>>>>> 26 jan. 2018 kl. 18:09 skrev Reed Dier <reed.d...@focusvq.com >>>>>> <mailto:reed.d...@focusvq.com>>: >>>>>> >>>>>> This is the exact issue that I ran into when starting my bluestore >>>>>> conversion journey. >>>>>> >>>>>> See my thread here: >>>>>> https://www.spinics.net/lists/ceph-users/msg41802.html >>>>>> >>>>>> Specifying --osd-id causes it to fail. >>>>>> >>>>>> Below are my steps for OSD replace/migrate from filestore to bluestore. >>>>>> >>>>>> BIG caveat here in that I am doing destructive replacement, in that I am >>>>>> not allowing my objects to be migrated off of the OSD I’m replacing >>>>>> before nuking it. >>>>>> With 8TB drives it just takes way too long, and I trust my failure >>>>>> domains and other hardware to get me through the backfills. >>>>>> So instead of 1) reading data off, writing data elsewhere 2) >>>>>> remove/re-add 3) reading data elsewhere, writing back on, I am taking >>>>>> step one out, and trusting my two other copies of the objects. Just >>>>>> wanted to clarify my steps. >>>>>> >>>>>> I also set norecover and norebalance flags immediately prior to running >>>>>> these commands so that it doesn’t try to start moving data >>>>>> unnecessarily. Then when done, remove those flags, and let it backfill. >>>>>> >>>>>>> systemctl stop ceph-osd@$ID.service <mailto:ceph-osd@$id.service> >>>>>>> ceph-osd -i $ID --flush-journal >>>>>>> umount /var/lib/ceph/osd/ceph-$ID >>>>>>> ceph-volume lvm zap /dev/$ID >>>>>>> ceph osd crush remove osd.$ID >>>>>>> ceph auth del osd.$ID >>>>>>> ceph osd rm osd.$ID >>>>>>> ceph-volume lvm create --bluestore --data /dev/$DATA --block.db >>>>>>> /dev/$NVME >>>>>> >>>>>> So essentially I fully remove the OSD from crush and the osdmap, and >>>>>> when I add the OSD back, like I would a new OSD, it fills in the numeric >>>>>> gap with the $ID it had before. >>>>>> >>>>>> Hope this is helpful. >>>>>> Been working well for me so far, doing 3 OSDs at a time (half of a >>>>>> failure domain). >>>>>> >>>>>> Reed >>>>>> >>>>>>> On Jan 26, 2018, at 10:01 AM, David <da...@visions.se >>>>>>> <mailto:da...@visions.se>> wrote: >>>>>>> >>>>>>> >>>>>>> Hi! >>>>>>> >>>>>>> On luminous 12.2.2 >>>>>>> >>>>>>> I'm migrating some OSDs from filestore to bluestore using the "simple" >>>>>>> method as described in docs: >>>>>>> http://docs.ceph.com/docs/master/rados/operations/bluestore-migration/#convert-existing-osds >>>>>>> Mark out and Replace. >>>>>>> >>>>>>> However, at 9.: ceph-volume create --bluestore --data $DEVICE --osd-id >>>>>>> $ID >>>>>>> it seems to create the bluestore but it fails to authenticate with the >>>>>>> old osd-id auth. >>>>>>> (the command above is also missing lvm or simple) >>>>>>> >>>>>>> I think it's related to this: >>>>>>> http://tracker.ceph.com/issues/22642 >>>>>>> >>>>>>> # ceph-volume lvm create --bluestore --data /dev/sdc --osd-id 0 >>>>>>> Running command: sudo vgcreate --force --yes >>>>>>> ceph-efad7df8-721d-43d8-8d02-449406e70b90 /dev/sdc >>>>>>> stderr: WARNING: lvmetad is running but disabled. Restart lvmetad >>>>>>> before enabling it! >>>>>>> stdout: Physical volume "/dev/sdc" successfully created >>>>>>> stdout: Volume group "ceph-efad7df8-721d-43d8-8d02-449406e70b90" >>>>>>> successfully created >>>>>>> Running command: sudo lvcreate --yes -l 100%FREE -n >>>>>>> osd-block-138ce507-f28a-45bf-814c-7fa124a9d9b9 >>>>>>> ceph-efad7df8-721d-43d8-8d02-449406e70b90 >>>>>>> stderr: WARNING: lvmetad is running but disabled. Restart lvmetad >>>>>>> before enabling it! >>>>>>> stdout: Logical volume >>>>>>> "osd-block-138ce507-f28a-45bf-814c-7fa124a9d9b9" created. >>>>>>> Running command: sudo mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-0 >>>>>>> Running command: chown -R ceph:ceph /dev/dm-4 >>>>>>> Running command: sudo ln -s >>>>>>> /dev/ceph-efad7df8-721d-43d8-8d02-449406e70b90/osd-block-138ce507-f28a-45bf-814c-7fa124a9d9b9 >>>>>>> /var/lib/ceph/osd/ceph-0/block >>>>>>> Running command: sudo ceph --cluster ceph --name client.bootstrap-osd >>>>>>> --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o >>>>>>> /var/lib/ceph/osd/ceph-0/activate.monmap >>>>>>> stderr: got monmap epoch 2 >>>>>>> Running command: ceph-authtool /var/lib/ceph/osd/ceph-0/keyring >>>>>>> --create-keyring --name osd.0 --add-key XXXXXXXX >>>>>>> stdout: creating /var/lib/ceph/osd/ceph-0/keyring >>>>>>> stdout: added entity osd.0 auth auth(auid = 18446744073709551615 key= >>>>>>> XXXXXXXX with 0 caps) >>>>>>> Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/keyring >>>>>>> Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/ >>>>>>> Running command: sudo ceph-osd --cluster ceph --osd-objectstore >>>>>>> bluestore --mkfs -i 0 --monmap /var/lib/ceph/osd/ceph-0/activate.monmap >>>>>>> --key **************************************** --osd-data >>>>>>> /var/lib/ceph/osd/ceph-0/ --osd-uuid >>>>>>> 138ce507-f28a-45bf-814c-7fa124a9d9b9 --setuser ceph --setgroup ceph >>>>>>> stderr: 2018-01-26 14:59:10.039549 7fd7ef951cc0 -1 >>>>>>> bluestore(/var/lib/ceph/osd/ceph-0//block) _read_bdev_label unable to >>>>>>> decode label at offset 102: buffer::malformed_input: void >>>>>>> bluestore_bdev_label_t::decode(ceph::buffer::list::iterator&) decode >>>>>>> past end of struct encoding >>>>>>> stderr: 2018-01-26 14:59:10.039744 7fd7ef951cc0 -1 >>>>>>> bluestore(/var/lib/ceph/osd/ceph-0//block) _read_bdev_label unable to >>>>>>> decode label at offset 102: buffer::malformed_input: void >>>>>>> bluestore_bdev_label_t::decode(ceph::buffer::list::iterator&) decode >>>>>>> past end of struct encoding >>>>>>> stderr: 2018-01-26 14:59:10.039925 7fd7ef951cc0 -1 >>>>>>> bluestore(/var/lib/ceph/osd/ceph-0//block) _read_bdev_label unable to >>>>>>> decode label at offset 102: buffer::malformed_input: void >>>>>>> bluestore_bdev_label_t::decode(ceph::buffer::list::iterator&) decode >>>>>>> past end of struct encoding >>>>>>> stderr: 2018-01-26 14:59:10.039984 7fd7ef951cc0 -1 >>>>>>> bluestore(/var/lib/ceph/osd/ceph-0/) _read_fsid unparsable uuid >>>>>>> stderr: 2018-01-26 14:59:11.359951 7fd7ef951cc0 -1 key XXXXXXXX >>>>>>> stderr: 2018-01-26 14:59:11.888476 7fd7ef951cc0 -1 created object >>>>>>> store /var/lib/ceph/osd/ceph-0/ for osd.0 fsid >>>>>>> efad7df8-721d-43d8-8d02-449406e70b90 >>>>>>> Running command: sudo ceph-bluestore-tool --cluster=ceph prime-osd-dir >>>>>>> --dev >>>>>>> /dev/ceph-efad7df8-721d-43d8-8d02-449406e70b90/osd-block-138ce507-f28a-45bf-814c-7fa124a9d9b9 >>>>>>> --path /var/lib/ceph/osd/ceph-0 >>>>>>> Running command: sudo ln -snf >>>>>>> /dev/ceph-efad7df8-721d-43d8-8d02-449406e70b90/osd-block-138ce507-f28a-45bf-814c-7fa124a9d9b9 >>>>>>> /var/lib/ceph/osd/ceph-0/block >>>>>>> Running command: chown -R ceph:ceph /dev/dm-4 >>>>>>> Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-0 >>>>>>> Running command: sudo systemctl enable >>>>>>> ceph-volume@lvm-0-138ce507-f28a-45bf-814c-7fa124a9d9b9 >>>>>>> stderr: Created symlink from >>>>>>> /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-0-138ce507-f28a-45bf-814c-7fa124a9d9b9.service >>>>>>> >>>>>>> <mailto:etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-0-138ce507-f28a-45bf-814c-7fa124a9d9b9.service> >>>>>>> to /lib/systemd/system/ceph-volume@.service. >>>>>>> Running command: sudo systemctl start ceph-osd@0 >>>>>>> >>>>>>> ceph-osd.0.log shows: >>>>>>> >>>>>>> 2018-01-26 15:09:07.379039 7f545d3b9cc0 4 rocksdb: >>>>>>> [/build/ceph-12.2.2/src/rocksdb/db/version_set.cc:2859 >>>>>>> <http://version_set.cc:2859/>] Recovered from manifest >>>>>>> file:db/MANIFEST-000095 succeeded,manifest_file_number is 95, >>>>>>> next_file_number is 97, last_sequence is 21, log_number is >>>>>>> 0,prev_log_number is 0,max_column_family is 0 >>>>>>> >>>>>>> 2018-01-26 15:09:07.379046 7f545d3b9cc0 4 rocksdb: >>>>>>> [/build/ceph-12.2.2/src/rocksdb/db/version_set.cc:2867 >>>>>>> <http://version_set.cc:2867/>] Column family [default] (ID 0), log >>>>>>> number is 94 >>>>>>> >>>>>>> 2018-01-26 15:09:07.379087 7f545d3b9cc0 4 rocksdb: EVENT_LOG_v1 >>>>>>> {"time_micros": 1516979347379083, "job": 1, "event": >>>>>>> "recovery_started", "log_files": [96]} >>>>>>> 2018-01-26 15:09:07.379091 7f545d3b9cc0 4 rocksdb: >>>>>>> [/build/ceph-12.2.2/src/rocksdb/db/db_impl_open.cc:482 >>>>>>> <http://db_impl_open.cc:482/>] Recovering log #96 mode 0 >>>>>>> 2018-01-26 15:09:07.379102 7f545d3b9cc0 4 rocksdb: >>>>>>> [/build/ceph-12.2.2/src/rocksdb/db/version_set.cc:2395 >>>>>>> <http://version_set.cc:2395/>] Creating manifest 98 >>>>>>> >>>>>>> 2018-01-26 15:09:07.380466 7f545d3b9cc0 4 rocksdb: EVENT_LOG_v1 >>>>>>> {"time_micros": 1516979347380463, "job": 1, "event": >>>>>>> "recovery_finished"} >>>>>>> 2018-01-26 15:09:07.381331 7f545d3b9cc0 4 rocksdb: >>>>>>> [/build/ceph-12.2.2/src/rocksdb/db/db_impl_open.cc:1063 >>>>>>> <http://db_impl_open.cc:1063/>] DB pointer 0x556ecb8c3000 >>>>>>> 2018-01-26 15:09:07.381353 7f545d3b9cc0 1 >>>>>>> bluestore(/var/lib/ceph/osd/ceph-0) _open_db opened rocksdb path db >>>>>>> options >>>>>>> compression=kNoCompression,max_write_buffer_number=4,min_write_buffer_number_to_merge=1,recycle_log_file_num=4,write_buffer_size=268435456,writable_file_max_buffer_size=0,compaction_readahead_size=2097152 >>>>>>> 2018-01-26 15:09:07.381616 7f545d3b9cc0 1 freelist init >>>>>>> 2018-01-26 15:09:07.381660 7f545d3b9cc0 1 >>>>>>> bluestore(/var/lib/ceph/osd/ceph-0) _open_alloc opening allocation >>>>>>> metadata >>>>>>> 2018-01-26 15:09:07.381679 7f545d3b9cc0 1 >>>>>>> bluestore(/var/lib/ceph/osd/ceph-0) _open_alloc loaded 447 G in 1 >>>>>>> extents >>>>>>> 2018-01-26 15:09:07.382077 7f545d3b9cc0 0 _get_class not permitted to >>>>>>> load kvs >>>>>>> 2018-01-26 15:09:07.382309 7f545d3b9cc0 0 <cls> >>>>>>> /build/ceph-12.2.2/src/cls/cephfs/cls_cephfs.cc:197 >>>>>>> <http://cls_cephfs.cc:197/>: loading cephfs >>>>>>> 2018-01-26 15:09:07.382583 7f545d3b9cc0 0 _get_class not permitted to >>>>>>> load sdk >>>>>>> 2018-01-26 15:09:07.382827 7f545d3b9cc0 0 <cls> >>>>>>> /build/ceph-12.2.2/src/cls/hello/cls_hello.cc:296 >>>>>>> <http://cls_hello.cc:296/>: loading cls_hello >>>>>>> 2018-01-26 15:09:07.385755 7f545d3b9cc0 0 _get_class not permitted to >>>>>>> load lua >>>>>>> 2018-01-26 15:09:07.386073 7f545d3b9cc0 0 osd.0 0 crush map has >>>>>>> features 288232575208783872, adjusting msgr requires for clients >>>>>>> 2018-01-26 15:09:07.386078 7f545d3b9cc0 0 osd.0 0 crush map has >>>>>>> features 288232575208783872 was 8705, adjusting msgr requires for mons >>>>>>> 2018-01-26 15:09:07.386079 7f545d3b9cc0 0 osd.0 0 crush map has >>>>>>> features 288232575208783872, adjusting msgr requires for osds >>>>>>> 2018-01-26 15:09:07.386132 7f545d3b9cc0 0 osd.0 0 load_pgs >>>>>>> 2018-01-26 15:09:07.386134 7f545d3b9cc0 0 osd.0 0 load_pgs opened 0 pgs >>>>>>> 2018-01-26 15:09:07.386137 7f545d3b9cc0 0 osd.0 0 using >>>>>>> weightedpriority op queue with priority op cut off at 64. >>>>>>> 2018-01-26 15:09:07.386580 7f545d3b9cc0 -1 osd.0 0 log_to_monitors >>>>>>> {default=true} >>>>>>> 2018-01-26 15:09:07.388077 7f545d3b9cc0 -1 osd.0 0 init authentication >>>>>>> failed: (1) Operation not permitted >>>>>>> >>>>>>> >>>>>>> The old osd is still there. >>>>>>> >>>>>>> # ceph osd tree >>>>>>> ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF >>>>>>> -1 2.60458 root default >>>>>>> -2 0.86819 host int1 >>>>>>> 0 ssd 0.43159 osd.0 destroyed 0 1.00000 >>>>>>> 3 ssd 0.43660 osd.3 up 1.00000 1.00000 >>>>>>> -3 0.86819 host int2 >>>>>>> 1 ssd 0.43159 osd.1 up 1.00000 1.00000 >>>>>>> 4 ssd 0.43660 osd.4 up 1.00000 1.00000 >>>>>>> -4 0.86819 host int3 >>>>>>> 2 ssd 0.43159 osd.2 up 1.00000 1.00000 >>>>>>> 5 ssd 0.43660 osd.5 up 1.00000 1.00000 >>>>>>> >>>>>>> >>>>>>> What's the best course of action? Purging osd.0, zapping the device >>>>>>> again and creating without --osd-id set? >>>>>>> >>>>>>> >>>>>>> Kind Regards, >>>>>>> >>>>>>> David Majchrzak >>>>>>> _______________________________________________ >>>>>>> ceph-users mailing list >>>>>>> ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com> >>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>>>> >>>>> >>>>> _______________________________________________ >>>>> ceph-users mailing list >>>>> ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com> >>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>> _______________________________________________ >>>> ceph-users mailing list >>>> ceph-users@lists.ceph.com >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>> _______________________________________________ >>> ceph-users mailing list >>> ceph-users@lists.ceph.com >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com