On Sat, Apr 14, 2018 at 5:17 PM, Rhian Resnick <rresn...@fau.edu> wrote:
> Afternoon, > > > Happily, I resolved this issue. > > > Running vgdisplay showed that ceph-volume tried to create a disk on failed > disk. (We didn't know we had a bad did so this is information that was new > to us) and when the command failed it left three bad volume groups. Since > you cannot rename them you need to use the following command to delete > them. > We actually tried to guard against this problem, but we failed in this specific situation where multiple VGs exist. I've created http://tracker.ceph.com/issues/23757 to follow up with this, thanks for reporting it! > > vgdisplay to find the bad volume groups > > vgremove --select vg_uuid=your uuid -f # -f forces it to be removed > > > Rhian Resnick > > Associate Director Middleware and HPC > > Office of Information Technology > > > Florida Atlantic University > > 777 Glades Road > <https://maps.google.com/?q=777+Glades+Road&entry=gmail&source=g>, CM22, > Rm 173B > > Boca Raton, FL 33431 > > Phone 561.297.2647 > > Fax 561.297.0222 > > [image: image] <https://hpc.fau.edu/wp-content/uploads/2015/01/image.jpg> > > > ------------------------------ > *From:* Rhian Resnick > *Sent:* Saturday, April 14, 2018 12:47 PM > *To:* Alfredo Deza > > *Cc:* ceph-users@lists.ceph.com > *Subject:* Re: [ceph-users] Error Creating OSD > > > Thanks all, > > > Here is a link to our our command being executed: https://pastebin. > com/iy8iSaKH > > > Here are the results from the command > > > Executed with debug enabled (after a zap with destroy) > > > [root@ceph-storage3 ~]# ceph-volume lvm create --bluestore --data > /dev/sdu > Running command: ceph-authtool --gen-print-key > Running command: ceph --cluster ceph --name client.bootstrap-osd --keyring > /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new > 664894a8-530a-4557-b2f4-1af5b391f2b7 > --> Was unable to complete a new OSD, will rollback changes > --> OSD will be fully purged from the cluster, because the ID was generated > Running command: ceph osd purge osd.140 --yes-i-really-mean-it > stderr: purged osd.140 > Traceback (most recent call last): > File "/sbin/ceph-volume", line 6, in <module> > main.Volume() > File "/usr/lib/python2.7/site-packages/ceph_volume/main.py", line 37, > in __init__ > self.main(self.argv) > File "/usr/lib/python2.7/site-packages/ceph_volume/decorators.py", line > 59, in newfunc > return f(*a, **kw) > File "/usr/lib/python2.7/site-packages/ceph_volume/main.py", line 153, > in main > terminal.dispatch(self.mapper, subcommand_args) > File "/usr/lib/python2.7/site-packages/ceph_volume/terminal.py", line > 182, in dispatch > instance.main() > File "/usr/lib/python2.7/site-packages/ceph_volume/devices/lvm/main.py", > line 38, in main > terminal.dispatch(self.mapper, self.argv) > File "/usr/lib/python2.7/site-packages/ceph_volume/terminal.py", line > 182, in dispatch > instance.main() > File "/usr/lib/python2.7/site-packages/ceph_volume/devices/lvm/create.py", > line 74, in main > self.create(args) > File "/usr/lib/python2.7/site-packages/ceph_volume/decorators.py", line > 16, in is_root > return func(*a, **kw) > File "/usr/lib/python2.7/site-packages/ceph_volume/devices/lvm/create.py", > line 26, in create > prepare_step.safe_prepare(args) > File "/usr/lib/python2.7/site-packages/ceph_volume/devices/lvm/prepare.py", > line 217, in safe_prepare > self.prepare(args) > File "/usr/lib/python2.7/site-packages/ceph_volume/decorators.py", line > 16, in is_root > return func(*a, **kw) > File "/usr/lib/python2.7/site-packages/ceph_volume/devices/lvm/prepare.py", > line 283, in prepare > block_lv = self.prepare_device(args.data, 'block', cluster_fsid, > osd_fsid) > File "/usr/lib/python2.7/site-packages/ceph_volume/devices/lvm/prepare.py", > line 193, in prepare_device > if api.get_vg(vg_name=vg_name): > File "/usr/lib/python2.7/site-packages/ceph_volume/api/lvm.py", line > 334, in get_vg > return vgs.get(vg_name=vg_name, vg_tags=vg_tags) > File "/usr/lib/python2.7/site-packages/ceph_volume/api/lvm.py", line > 429, in get > raise MultipleVGsError(vg_name) > ceph_volume.exceptions.MultipleVGsError: Got more than 1 result looking > for volume group: ceph-6a2e8f21-bca2-492b-8869-eecc995216cc > > > > Rhian Resnick > > Associate Director Middleware and HPC > > Office of Information Technology > > > Florida Atlantic University > > 777 Glades Road > <https://maps.google.com/?q=777+Glades+Road&entry=gmail&source=g>, CM22, > Rm 173B > > Boca Raton, FL 33431 > > Phone 561.297.2647 > > Fax 561.297.0222 > > [image: image] <https://hpc.fau.edu/wp-content/uploads/2015/01/image.jpg> > > > ------------------------------ > *From:* Alfredo Deza <ad...@redhat.com> > *Sent:* Saturday, April 14, 2018 8:45 AM > *To:* Rhian Resnick > *Cc:* ceph-users@lists.ceph.com > *Subject:* Re: [ceph-users] Error Creating OSD > > > > On Fri, Apr 13, 2018 at 8:20 PM, Rhian Resnick <rresn...@fau.edu> wrote: > > Evening, > > When attempting to create an OSD we receive the following error. > > [ceph-admin@ceph-storage3 ~]$ sudo ceph-volume lvm create --bluestore > --data /dev/sdu > Running command: ceph-authtool --gen-print-key > Running command: ceph --cluster ceph --name client.bootstrap-osd --keyring > /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new > c8cb8cff-dad9-48b8-8d77-6f130a4b629d > --> Was unable to complete a new OSD, will rollback changes > --> OSD will be fully purged from the cluster, because the ID was generated > Running command: ceph osd purge osd.140 --yes-i-really-mean-it > stderr: purged osd.140 > --> MultipleVGsError: Got more than 1 result looking for volume group: > ceph-6a2e8f21-bca2-492b-8869-eecc995216cc > > Any hints on what to do? This occurs when we attempt to create osd's on > this node. > > > Can you use a paste site and get the /var/log/ceph/ceph-volume.log > contents? Also, if you could try the same command but with: > > CEPH_VOLUME_DEBUG=1 > > I think you are hitting two issues here: > > 1) Somehow `osd new` is not completing and failing > 2) The `purge` command to wipe out the LV is getting multiple LV's and > cannot make sure to match the one it used. > > #2 definitely looks like something we are doing wrong, and #1 can have a > lot of different causes. The logs would be tremendously helpful! > > > Rhian Resnick > > Associate Director Middleware and HPC > > Office of Information Technology > > > Florida Atlantic University > > 777 Glades Road > <https://maps.google.com/?q=777+Glades+Road&entry=gmail&source=g>, CM22, > Rm 173B > > Boca Raton, FL 33431 > > Phone 561.297.2647 > > Fax 561.297.0222 > > [image: image] <https://hpc.fau.edu/wp-content/uploads/2015/01/image.jpg> > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com