Hello Fabrizio, On Thu, Jan 30, 2020 at 12:46:16PM +0100, Fabrizio Cuseo wrote: > > I have installed a new cluster with the last release, with a local ceph > storage. > I also have 2 old and smaller clusters, and I need to migrate all the VMs to > the new cluster. > The best method i have used in past is to add on the NEW cluster the RBD > storage of the old cluster, so I can stop the VM, move the .cfg file, start > the vm (all those operations are really quick), and move the disk (online) > from the old storage to the new storage. > > But now, if I add the RBD storage, copying the keyring file of the old > cluster to the new cluster, naming as the storage ID, and using the old > cluster monitors IP, i can see the storage summary (space total and used), > but when I go to "content", i have this error: "rbd error: rbd: listing > images failed: (95) Operation not supported (500)". > > If, from the new cluster CLI, i use the command: > > rbd -k /etc/pve/priv/ceph/CephOLD.keyring -m 172.16.20.31 ls rbd2 > > I can see the list of disk images, but also the error: "librbd::api::Trash: > list: error listing rbd trash entries: (95) Operation not supported" > > > The new cluster ceph release is Nautilus, and the old one is firefly. > > Some idea ? As said by others already, there is no direct way. Best OFC would be to do a backup + restore. But in any case, you will need a shared storage that can be reached by both clusters, eg. like NFS. And watch out, as one cluster can potentially destroy disks from the other cluster on the shared storage.
-- Cheers, Alwin _______________________________________________ pve-user mailing list pve-user@pve.proxmox.com https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user