Why are you making the same rbd to multiple servers? On Wed, Feb 27, 2019, 9:50 AM Ilya Dryomov <idryo...@gmail.com> wrote:
> On Wed, Feb 27, 2019 at 12:00 PM Thomas <74cmo...@gmail.com> wrote: > > > > Hi, > > I have noticed an error when writing to a mapped RBD. > > Therefore I unmounted the block device. > > Then I tried to unmap it w/o success: > > ld2110:~ # rbd unmap /dev/rbd0 > > rbd: sysfs write failed > > rbd: unmap failed: (16) Device or resource busy > > > > The same block device is mapped on another client and there are no > issues: > > root@ld4257:~# rbd info hdb-backup/ld2110 > > rbd image 'ld2110': > > size 7.81TiB in 2048000 objects > > order 22 (4MiB objects) > > block_name_prefix: rbd_data.3cda0d6b8b4567 > > format: 2 > > features: layering > > flags: > > create_timestamp: Fri Feb 15 10:53:50 2019 > > root@ld4257:~# rados -p hdb-backup listwatchers rbd_data.3cda0d6b8b4567 > > error listing watchers hdb-backup/rbd_data.3cda0d6b8b4567: (2) No such > > file or directory > > root@ld4257:~# rados -p hdb-backup listwatchers > rbd_header.3cda0d6b8b4567 > > watcher=10.76.177.185:0/1144812735 client.21865052 cookie=1 > > watcher=10.97.206.97:0/4023931980 client.18484780 > > cookie=18446462598732841027 > > > > > > Question: > > How can I force to unmap the RBD on client ld2110 (= 10.76.177.185)? > > Hi Thomas, > > It appears that /dev/rbd0 is still open on that node. > > Was the unmount successful? Which filesystem (ext4, xfs, etc)? > > What is the output of "ps aux | grep rbd" on that node? > > Try lsof, fuser, check for LVM volumes and multipath -- these have been > reported to cause this issue previously: > > http://tracker.ceph.com/issues/12763 > > Thanks, > > Ilya > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com