Hi Klimenko, I did a migration from filestore to bluestroe on centos7 with ceph version 12.2.5. As it's the pro environment, I removed and recreated OSDs on each server at a time, online. Although I migreated on centos, I create osd manually so that you can have a try.
Except one raid1 disk for system, my server has one SSD(sdg) for wal/db, and 5 SATAs(sdb-sdf) for storage. Here is my steps of replacing osd.16 and osd.17: 1. set osds out and remove them from cluster ceph osd out 16 && ceph osd out 17 systemctl stop ceph-osd@16 && systemctl stop ceph-osd@17 ceph osd crush remove osd.16 && ceph osd crush remove osd.17 ceph osd rm osd.16 && ceph osd rm osd.17 ps. if you replace few disks you don't have to wait until the recovery done as the cluster would heal itself. 2. clear the partition infomation on the disks sgdisk --zap-all /dev/sdb sgdisk --zap-all /dev/sdc 3. make meta partitions sgdisk --new=1:0:+1GB --change-name=1:osd_data_16 --partition-guid=1:$(uuidgen) --mbrtogpt -- /dev/sdb sgdisk --largest-new=2 --change-name=2:bluestore_block_16 --partition-guid=2:$(uuidgen) --mbrtogpt -- /dev/sdb sgdisk --new=1:0:+1GB --change-name=1:osd_data_17 --partition-guid=1:$(uuidgen) --mbrtogpt -- /dev/sdc sgdisk --largest-new=2 --change-name=2:bluestore_block_17 --partition-guid=2:$(uuidgen) --mbrtogpt -- /dev/sdc 4. format partitions mkfs -t xfs -f -i size=2048 -- /dev/sdb1 mkfs -t xfs -f -i size=2048 -- /dev/sdc1 5. make wal/db partitions sgdisk --new=1:0:+1GB --change-name=1:bluestore_block_db_16 --partition-guid=1:$(uuidgen) --mbrtogpt -- /dev/sdg sgdisk --new=2:0:+8GB --change-name=2:bluestore_block_wal_16 --partition-guid=2:$(uuidgen) --mbrtogpt -- /dev/sdg sgdisk --new=3:0:+1GB --change-name=3:bluestore_block_db_17 --partition-guid=3:$(uuidgen) --mbrtogpt -- /dev/sdg sgdisk --new=4:0:+8GB --change-name=4:bluestore_block_wal_17 --partition-guid=4:$(uuidgen) --mbrtogpt -- /dev/sdg 6. create osd ceph osd create 7. prepare the osd mount /dev/sdb1 /var/lib/ceph/osd/ceph-16 mount /dev/sdc1 /var/lib/ceph/osd/ceph-17 echo "bluestore" > /var/lib/ceph/osd/ceph-16/type echo "bluestore" > /var/lib/ceph/osd/ceph-17/type 8. edit ceph.conf [osd.16] host = ceph-osd1 osd data = /var/lib/ceph/osd/ceph-16/ bluestore block path = /dev/disk/by-partlabel/bluestore_block_16 bluestore block db path = /dev/disk/by-partlabel/bluestore_block_db_16 bluestore block wal path = /dev/disk/by-partlabel/bluestore_block_wal_16 [osd.17] host = ceph-osd1 osd data = /var/lib/ceph/osd/ceph-17/ bluestore block path = /dev/disk/by-partlabel/bluestore_block_17 bluestore block db path = /dev/disk/by-partlabel/bluestore_block_db_17 bluestore block wal path = /dev/disk/by-partlabel/bluestore_block_wal_17 9. make keys ceph-osd -i 16 --mkkey --mkfs ceph-osd -i 17 --mkkey --mkfs 10. authorize ceph auth add osd.16 osd 'allow *' mon 'allow profile osd' mgr 'allow profile osd' -i /var/lib/ceph/osd/ceph-16/keyring ceph auth add osd.17 osd 'allow *' mon 'allow profile osd' mgr 'allow profile osd' -i /var/lib/ceph/osd/ceph-17/keyring 11. edit crushmap ceph osd crush add 16 1 host=ceph-osd1 ceph osd crush add 17 1 host=ceph-osd1 12. start OSDs Last but not least, you may edit udev rules as well as chown ceph osd directories so that ceph has permisson on wiriting the disks and directories. I hope it helps. Thanks Klimenko, Roman <rklime...@itkey.com> 于2018年11月16日周五 上午8:35写道: > Hi everyone! > > As I noticed, ceph-volume lacks Ubuntu Trusty compatibility > https://tracker.ceph.com/issues/23496 > > So, I can't follow this instruction > http://docs.ceph.com/docs/mimic/rados/operations/bluestore-migration/ > > Do I have any other option to migrate my Filestore osds (Luminous 12.2.9) > to Bluestore? > > P.S This is a test environment, so I can try anything > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com