After removing the —osd-id flag, everything came up normally. > -2 21.82448 host node24 > 0 hdd 7.28450 osd.0 up 1.00000 1.00000 > 8 hdd 7.26999 osd.8 up 1.00000 1.00000 > 16 hdd 7.26999 osd.16 up 1.00000 1.00000
Given the vanilla-ness to this ceph-volume command, is this something ceph-deploy-able? I’m seeing ceph-deploy 1.5.39 as the latest stable release. > ceph-deploy --username root disk zap $NODE:$HDD > ceph-deploy --username root osd create $NODE:$HDD:$SSD In that example $HDD is the main OSD device, and $SSD is the NVMe partition I want to use for block.db (and block.wal). Or is the syntax different from the filestore days? And I am assuming that no --bluestore would be necessary given that I am reading that bluestore is the default and filestore requires intervention. Thanks, Reed > On Jan 9, 2018, at 2:10 PM, Reed Dier <reed.d...@focusvq.com> wrote: > >> -2 21.81000 host node24 >> 0 hdd 7.26999 osd.0 destroyed 0 >> 1.00000 >> 8 hdd 7.26999 osd.8 up 1.00000 >> 1.00000 >> 16 hdd 7.26999 osd.16 up 1.00000 >> 1.00000 > > Should I do these prior to running without the osd-id specified? >> # ceph osd crush remove osd.$ID >> # ceph auth del osd.$ID >> # ceph osd rm osd.$ID > > > And then it fill in the missing osd.0. > Will set norebalance flag first to prevent data reshuffle upon the osd being > removed from the crush map. > > Thanks, > > Reed > >> On Jan 9, 2018, at 2:05 PM, Alfredo Deza <ad...@redhat.com >> <mailto:ad...@redhat.com>> wrote: >> >> On Tue, Jan 9, 2018 at 2:19 PM, Reed Dier <reed.d...@focusvq.com >> <mailto:reed.d...@focusvq.com>> wrote: >>> Hi ceph-users, >>> >>> Hoping that this is something small that I am overlooking, but could use the >>> group mind to help. >>> >>> Ceph 12.2.2, Ubuntu 16.04 environment. >>> OSD (0) is an 8TB spinner (/dev/sda) and I am moving from a filestore >>> journal to a blocks.db and WAL device on an NVMe partition (/dev/nvme0n1p5). >>> >>> I have an OSD that I am trying to convert to bluestore and running into some >>> trouble. >>> >>> Started here until the ceps-volume create statement, which doesn’t work. >>> http://docs.ceph.com/docs/master/rados/operations/bluestore-migration/ >>> <http://docs.ceph.com/docs/master/rados/operations/bluestore-migration/> >>> Worth mentioning I also flushed the journal on the nvme partition before >>> nuking the OSD. >>> >>> $ sudo ceph-osd -i 0 --flush-journal >>> >>> >>> So I first started with this command: >>> >>> $ sudo ceph-volume lvm create --bluestore --data /dev/sda --block.db >>> /dev/nvme0n1p5 --osd-id 0 >>> >>> >>> Pastebin to the ceph-volume log: https://pastebin.com/epkM3aP6 >>> >>> However the OSD doesn’t start. >> >> I was just able to replicate this by using an ID that doesn't exist in >> the cluster. On a cluster with just one OSD (with an ID of 0) I >> created >> an OSD with --osd-id 3, and had the exact same results. >> >>> >>> Pastebin to ceph-osd log: https://pastebin.com/9qEsAJzA >>> <https://pastebin.com/9qEsAJzA> >>> >>> I tried restarting the process, by deleting the LVM structures, zapping the >>> disk using ceph-volume. >>> This time using prepare and activate instead of create. >>> >>> $ sudo ceph-volume lvm prepare --bluestore --data /dev/sda --block.db >>> /dev/nvme0n1p5 --osd-id 0 >>> >>> $ sudo ceph-volume lvm activate --bluestore 0 >>> 227e1721-cd2e-4d7e-bb48-bc2bb715a038 >>> >>> >>> Also ran the enable on the ceph-volume systemd unit per >>> http://docs.ceph.com/docs/master/install/manual-deployment/ >>> <http://docs.ceph.com/docs/master/install/manual-deployment/> >>> >>> $ sudo systemctl enable >>> ceph-volume@lvm-0-227e1721-cd2e-4d7e-bb48-bc2bb715a038 >>> >>> >>> Same results. >>> >>> Any help is greatly appreciated. >> >> Could you try without passing --osd-id ? >>> >>> Thanks, >>> >>> Reed >>> >>> _______________________________________________ >>> ceph-users mailing list >>> ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com> >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>> <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com