When cloning the snapshot on the remote cluster I can't see my ext4 filesystem. 

Using the same exact snapshot on both sides. Shouldn't this be consistent? 

Primary Site 
root@Ccscephtest1:~# rbd snap ls --all CephTestPool1/vm-100-disk-0 | grep 
TestSnapper1 
10621 TestSnapper1 2 TiB Thu Jan 21 08:15:22 2021 user 

root@Ccscephtest1:~# rbd clone CephTestPool1/vm-100-disk-0@TestSnapper1 
CephTestPool1/vm-100-disk-0-CLONE 
root@Ccscephtest1:~# rbd-nbd map CephTestPool1/vm-100-disk-0-CLONE --id admin 
--keyring /etc/ceph/ceph.client.admin.keyring 
/dev/nbd0 
root@Ccscephtest1:~# mount /dev/nbd0 /usr2 

Secondary Site 
root@Bunkcephtest1:~# rbd snap ls --all CephTestPool1/vm-100-disk-0 | grep 
TestSnapper1 
10430 TestSnapper1 2 TiB Thu Jan 21 08:20:08 2021 user 

root@Bunkcephtest1:~# rbd clone CephTestPool1/vm-100-disk-0@TestSnapper1 
CephTestPool1/vm-100-disk-0-CLONE 
root@Bunkcephtest1:~# rbd-nbd map CephTestPool1/vm-100-disk-0-CLONE --id admin 
--keyring /etc/ceph/ceph.client.admin.keyring 
/dev/nbd0 
root@Bunkcephtest1:~# mount /dev/nbd0 /usr2 
mount: /usr2: wrong fs type, bad option, bad superblock on /dev/nbd0, missing 
codepage or helper program, or other error. 




From: "adamb" <ad...@medent.com> 
To: "dillaman" <dilla...@redhat.com> 
Cc: "Eugen Block" <ebl...@nde.ag>, "ceph-users" <ceph-users@ceph.io>, "Matt 
Wilder" <matt.wil...@bitmex.com> 
Sent: Wednesday, January 20, 2021 3:42:46 PM 
Subject: Re: [ceph-users] Re: RBD-Mirror Snapshot Backup Image Uses 

Awesome information. I new I had to be missing something. 

All of my clients will be far newer than mimic so I don't think that will be an 
issue. 

Added the following to my ceph.conf on both clusters. 

rbd_default_clone_format = 2 

root@Bunkcephmon2:~# rbd clone CephTestPool1/vm-100-disk-0@TestSnapper1 
CephTestPool2/vm-100-disk-0-CLONE 
root@Bunkcephmon2:~# rbd ls CephTestPool2 
vm-100-disk-0-CLONE 

I am sure I will be back with more questions. Hoping to replace our Nimble 
storage with Ceph and NVMe. 

Appreciate it! 


From: "Jason Dillaman" <jdill...@redhat.com> 
To: "adamb" <ad...@medent.com> 
Cc: "Eugen Block" <ebl...@nde.ag>, "ceph-users" <ceph-users@ceph.io>, "Matt 
Wilder" <matt.wil...@bitmex.com> 
Sent: Wednesday, January 20, 2021 3:28:39 PM 
Subject: Re: [ceph-users] Re: RBD-Mirror Snapshot Backup Image Uses 

On Wed, Jan 20, 2021 at 3:10 PM Adam Boyhan <ad...@medent.com> wrote: 
> 
> That's what I though as well, specially based on this. 
> 
> 
> 
> Note 
> 
> You may clone a snapshot from one pool to an image in another pool. For 
> example, you may maintain read-only images and snapshots as templates in one 
> pool, and writeable clones in another pool. 
> 
> root@Bunkcephmon2:~# rbd clone CephTestPool1/vm-100-disk-0@TestSnapper1 
> CephTestPool2/vm-100-disk-0-CLONE 
> 2021-01-20T15:06:35.854-0500 7fb889ffb700 -1 librbd::image::CloneRequest: 
> 0x55c7cf8417f0 validate_parent: parent snapshot must be protected 
> 
> root@Bunkcephmon2:~# rbd snap protect 
> CephTestPool1/vm-100-disk-0@TestSnapper1 
> rbd: protecting snap failed: (30) Read-only file system 

You have two options: (1) protect the snapshot on the primary image so 
that the protection status replicates or (2) utilize RBD clone v2 
which doesn't require protection but does require Mimic or later 
clients [1]. 

> 
> From: "Eugen Block" <ebl...@nde.ag> 
> To: "adamb" <ad...@medent.com> 
> Cc: "ceph-users" <ceph-users@ceph.io>, "Matt Wilder" <matt.wil...@bitmex.com> 
> Sent: Wednesday, January 20, 2021 3:00:54 PM 
> Subject: Re: [ceph-users] Re: RBD-Mirror Snapshot Backup Image Uses 
> 
> But you should be able to clone the mirrored snapshot on the remote 
> cluster even though it’s not protected, IIRC. 
> 
> 
> Zitat von Adam Boyhan <ad...@medent.com>: 
> 
> > Two separate 4 node clusters with 10 OSD's in each node. Micron 9300 
> > NVMe's are the OSD drives. Heavily based on the Micron/Supermicro 
> > white papers. 
> > 
> > When I attempt to protect the snapshot on a remote image, it errors 
> > with read only. 
> > 
> > root@Bunkcephmon2:~# rbd snap protect 
> > CephTestPool1/vm-100-disk-0@TestSnapper1 
> > rbd: protecting snap failed: (30) Read-only file system 
> > _______________________________________________ 
> > ceph-users mailing list -- ceph-users@ceph.io 
> > To unsubscribe send an email to ceph-users-le...@ceph.io 
> _______________________________________________ 
> ceph-users mailing list -- ceph-users@ceph.io 
> To unsubscribe send an email to ceph-users-le...@ceph.io 

[1] https://ceph.io/community/new-mimic-simplified-rbd-image-cloning/ 

-- 
Jason 

_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to