The rbd_directory object is empty -- all data is stored as omap key/value pairs which you can list via "rados listomapvals rbd_directory". What is the output when you run "rbd ls --debug-ms=1 glebe-sata" and "rbd info --debug-ms=1 glebe-sata/hypervtst-lun04"? I am interested in the lines that looks like the following:
** rbd ls ** 2016-06-05 22:22:54.816801 7f25d4e4d1c0 1 -- 127.0.0.1:0/2033136975 --> 127.0.0.1:6800/29402 -- osd_op(client.4111.0:2 0.30a98c1c rbd_directory [call rbd.dir_list] snapc 0=[] ack+read+known_if_redirected e7) v7 -- ?+0 0x5598b0459410 con 0x5598b04580d0 2016-06-05 22:22:54.817396 7f25b8207700 1 -- 127.0.0.1:0/2033136975 <== osd.0 127.0.0.1:6800/29402 2 ==== osd_op_reply(2 rbd_directory [call] v0'0 uv1 ondisk = 0) v7 ==== 133+0+27 (2231830616 0 2896097477) 0x7f258c000a20 con 0x5598b04580d0foo ** rbd info ** 2016-06-05 22:25:54.534064 7fab3cff9700 1 -- 127.0.0.1:0/951637948 --> 127.0.0.1:6800/29402 -- osd_op(client.4112.0:2 0.6a181655 rbd_id.foo [call rbd.get_id] snapc 0=[] ack+read+known_if_redirected e7) v7 -- ?+0 0x7fab180020a0 con 0x55e833b5e520 2016-06-05 22:25:54.534434 7fab4c589700 1 -- 127.0.0.1:0/951637948 <== osd.0 127.0.0.1:6800/29402 2 ==== osd_op_reply(2 rbd_id.foo [call] v0'0 uv2 ondisk = 0) v7 ==== 130+0+16 (2464064221 0 855464132) 0x7fab24000b40 con 0x55e833b5e520 I suspect you are having issues with executing OSD class methods for some reason (like rbd.dir_list against rbd_directory and rbd.get_id against rbd_id.<image name>). On Sun, Jun 5, 2016 at 9:16 PM, Adrian Saul <adrian.s...@tpgtelecom.com.au> wrote: > > Seems like my rbd_directory is empty for some reason: > > [root@ceph-glb-fec-02 ceph]# rados get -p glebe-sata rbd_directory /tmp/dir > [root@ceph-glb-fec-02 ceph]# strings /tmp/dir > [root@ceph-glb-fec-02 ceph]# ls -la /tmp/dir > -rw-r--r--. 1 root root 0 Jun 6 11:12 /tmp/dir > > [root@ceph-glb-fec-02 ceph]# rados stat -p glebe-sata rbd_directory > glebe-sata/rbd_directory mtime 2016-06-06 10:18:28.000000, size 0 > > > >> -----Original Message----- >> From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of >> Adrian Saul >> Sent: Monday, 6 June 2016 11:11 AM >> To: dilla...@redhat.com >> Cc: ceph-users@lists.ceph.com >> Subject: Re: [ceph-users] Jewel upgrade - rbd errors after upgrade >> >> >> No - it throws a usage error - if I add a file argument after it works: >> >> [root@ceph-glb-fec-02 ceph]# rados -p glebe-sata get rbd_id.hypervtst- >> lun04 /tmp/crap >> [root@ceph-glb-fec-02 ceph]# cat /tmp/crap 109eb01f5f89de >> >> stat works: >> >> [root@ceph-glb-fec-02 ceph]# rados -p glebe-sata stat rbd_id.hypervtst- >> lun04 >> glebe-sata/rbd_id.hypervtst-lun04 mtime 2016-06-06 10:55:08.000000, size 18 >> >> >> I can do a rados ls: >> >> [root@ceph-glb-fec-02 ceph]# rados ls -p glebe-sata|grep rbd_id >> rbd_id.cloud2sql-lun01 >> rbd_id.glbcluster3-vm17 >> rbd_id.holder <<< a create that said it failed while I was debugging this >> rbd_id.pvtcloud-nfs01 >> rbd_id.hypervtst-lun05 >> rbd_id.test02 >> rbd_id.cloud2sql-lun02 >> rbd_id.fiotest2 >> rbd_id.radmast02-lun04 >> rbd_id.hypervtst-lun04 >> rbd_id.cloud2fs-lun00 >> rbd_id.radmast02-lun03 >> rbd_id.hypervtst-lun00 >> rbd_id.cloud2sql-lun00 >> rbd_id.radmast02-lun02 >> >> >> > -----Original Message----- >> > From: Jason Dillaman [mailto:jdill...@redhat.com] >> > Sent: Monday, 6 June 2016 11:00 AM >> > To: Adrian Saul >> > Cc: ceph-users@lists.ceph.com >> > Subject: Re: [ceph-users] Jewel upgrade - rbd errors after upgrade >> > >> > Are you able to successfully run the following command successfully? >> > >> > rados -p glebe-sata get rbd_id.hypervtst-lun04 >> > >> > >> > >> > On Sun, Jun 5, 2016 at 8:49 PM, Adrian Saul >> > <adrian.s...@tpgtelecom.com.au> wrote: >> > > >> > > I upgraded my Infernalis semi-production cluster to Jewel on Friday. >> > > While >> > the upgrade went through smoothly (aside from a time wasting >> > restorecon /var/lib/ceph in the selinux package upgrade) and the >> > services continued running without interruption. However this morning >> > when I went to create some new RBD images I am unable to do much at all >> with RBD. >> > > >> > > Just about any rbd command fails with an I/O error. I can run >> > showmapped but that is about it - anything like an ls, info or status >> > fails. This applies to all my pools. >> > > >> > > I can see no errors in any log files that appear to suggest an >> > > issue. I have >> > also tried the commands on other cluster members that have not done >> > anything with RBD before (I was wondering if perhaps the kernel rbd >> > was pinning the old library version open or something) but the same error >> occurs. >> > > >> > > Where can I start trying to resolve this? >> > > >> > > Cheers, >> > > Adrian >> > > >> > > >> > > [root@ceph-glb-fec-01 ceph]# rbd ls glebe-sata >> > > rbd: list: (5) Input/output error >> > > 2016-06-06 10:41:31.792720 7f53c06a2d80 -1 librbd: error listing >> > > image in directory: (5) Input/output error >> > > 2016-06-06 10:41:31.792749 7f53c06a2d80 -1 librbd: error listing v2 >> > > images: (5) Input/output error >> > > >> > > [root@ceph-glb-fec-01 ceph]# rbd ls glebe-ssd >> > > rbd: list: (5) Input/output error >> > > 2016-06-06 10:41:33.956648 7f90de663d80 -1 librbd: error listing >> > > image in directory: (5) Input/output error >> > > 2016-06-06 10:41:33.956672 7f90de663d80 -1 librbd: error listing v2 >> > > images: (5) Input/output error >> > > >> > > [root@ceph-glb-fec-02 ~]# rbd showmapped >> > > id pool image snap device >> > > 0 glebe-sata test02 - /dev/rbd0 >> > > 1 glebe-ssd zfstest - /dev/rbd1 >> > > 10 glebe-sata hypervtst-lun00 - /dev/rbd10 >> > > 11 glebe-sata hypervtst-lun02 - /dev/rbd11 >> > > 12 glebe-sata hypervtst-lun03 - /dev/rbd12 >> > > 13 glebe-ssd nspprd01_lun00 - /dev/rbd13 >> > > 14 glebe-sata cirrux-nfs01 - /dev/rbd14 >> > > 15 glebe-sata hypervtst-lun04 - /dev/rbd15 >> > > 16 glebe-sata hypervtst-lun05 - /dev/rbd16 >> > > 17 glebe-sata pvtcloud-nfs01 - /dev/rbd17 >> > > 18 glebe-sata cloud2sql-lun00 - /dev/rbd18 >> > > 19 glebe-sata cloud2sql-lun01 - /dev/rbd19 >> > > 2 glebe-sata radmast02-lun00 - /dev/rbd2 >> > > 20 glebe-sata cloud2sql-lun02 - /dev/rbd20 >> > > 21 glebe-sata cloud2fs-lun00 - /dev/rbd21 >> > > 22 glebe-sata cloud2fs-lun01 - /dev/rbd22 >> > > 3 glebe-sata radmast02-lun01 - /dev/rbd3 >> > > 4 glebe-sata radmast02-lun02 - /dev/rbd4 >> > > 5 glebe-sata radmast02-lun03 - /dev/rbd5 >> > > 6 glebe-sata radmast02-lun04 - /dev/rbd6 >> > > 7 glebe-ssd sybase_iquser02_lun00 - /dev/rbd7 >> > > 8 glebe-ssd sybase_iquser03_lun00 - /dev/rbd8 >> > > 9 glebe-ssd sybase_iquser04_lun00 - /dev/rbd9 >> > > >> > > [root@ceph-glb-fec-02 ~]# rbd status glebe-sata/hypervtst-lun04 >> > > 2016-06-06 10:47:30.221453 7fc0030dc700 -1 librbd::image::OpenRequest: >> > > failed to retrieve image id: (5) Input/output error >> > > 2016-06-06 10:47:30.221556 7fc0028db700 -1 librbd::ImageState: >> > > failed to open image: (5) Input/output error >> > > rbd: error opening image hypervtst-lun04: (5) Input/output error >> > > Confidentiality: This email and any attachments are confidential and >> > > may be >> > subject to copyright, legal or some other professional privilege. They >> > are intended solely for the attention and use of the named >> > addressee(s). They may only be copied, distributed or disclosed with >> > the consent of the copyright owner. If you have received this email by >> > mistake or by breach of the confidentiality clause, please notify the >> > sender immediately by return email and delete or destroy all copies of >> > the email. Any confidentiality, privilege or copyright is not waived >> > or lost because this email has been sent to you by mistake. >> > > _______________________________________________ >> > > ceph-users mailing list >> > > ceph-users@lists.ceph.com >> > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > >> > >> > >> > -- >> > Jason >> Confidentiality: This email and any attachments are confidential and may be >> subject to copyright, legal or some other professional privilege. They are >> intended solely for the attention and use of the named addressee(s). They >> may only be copied, distributed or disclosed with the consent of the >> copyright owner. If you have received this email by mistake or by breach of >> the confidentiality clause, please notify the sender immediately by return >> email and delete or destroy all copies of the email. Any confidentiality, >> privilege or copyright is not waived or lost because this email has been sent >> to you by mistake. >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > Confidentiality: This email and any attachments are confidential and may be > subject to copyright, legal or some other professional privilege. They are > intended solely for the attention and use of the named addressee(s). They may > only be copied, distributed or disclosed with the consent of the copyright > owner. If you have received this email by mistake or by breach of the > confidentiality clause, please notify the sender immediately by return email > and delete or destroy all copies of the email. Any confidentiality, privilege > or copyright is not waived or lost because this email has been sent to you by > mistake. -- Jason _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com