Re: [ceph-users] Centos 7 OSD silently fail to start
We use ceph-disk without any issues on CentOS7. If you want to do a manual deployment, verfiy you aren't missing any steps in http://ceph.com/docs/master/install/manual-deployment/#long-form. On Tue, Feb 24, 2015 at 5:46 PM, Barclay Jameson wrote: > I have tried to install ceph using ceph-deploy but sgdisk seems to > have too many issues so I did a manual install. After mkfs.btrfs on > the disks and journals and mounted them I then tried to start the osds > which failed. The first error was: > #/etc/init.d/ceph start osd.0 > /etc/init.d/ceph: osd.0 not found (/etc/ceph/ceph.conf defines , > /var/lib/ceph defines ) > > I then manually added the osds to the conf file with the following as > an example: > [osd.0] > osd_host = node01 > > Now when I run the command : > # /etc/init.d/ceph start osd.0 > > There is no error or output from the command and in fact when I do a > ceph -s no osds are listed as being up. > Doing as ps aux | grep -i ceph or ps aux | grep -i osd shows there are > no osd running. > I also have done htop to see if any process are running and none are shown. > > I had this working on SL6.5 with Firefly but Giant on Centos 7 has > been nothing but a giant pain. > ___ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Centos 7 OSD silently fail to start
Check firewall rules and selinux. It sometimes is a pain in the ... :) 25 lut 2015 01:46 "Barclay Jameson" napisał(a): > I have tried to install ceph using ceph-deploy but sgdisk seems to > have too many issues so I did a manual install. After mkfs.btrfs on > the disks and journals and mounted them I then tried to start the osds > which failed. The first error was: > #/etc/init.d/ceph start osd.0 > /etc/init.d/ceph: osd.0 not found (/etc/ceph/ceph.conf defines , > /var/lib/ceph defines ) > > I then manually added the osds to the conf file with the following as > an example: > [osd.0] > osd_host = node01 > > Now when I run the command : > # /etc/init.d/ceph start osd.0 > > There is no error or output from the command and in fact when I do a > ceph -s no osds are listed as being up. > Doing as ps aux | grep -i ceph or ps aux | grep -i osd shows there are > no osd running. > I also have done htop to see if any process are running and none are shown. > > I had this working on SL6.5 with Firefly but Giant on Centos 7 has > been nothing but a giant pain. > ___ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Centos 7 OSD silently fail to start
Also, did you successfully start your monitor(s), and define/create the OSDs within the Ceph cluster itself? There are several steps to creating a Ceph cluster manually. I'm unsure if you have done the steps to actually create and register the OSDs with the cluster. - Travis On Wed, Feb 25, 2015 at 9:49 AM, Leszek Master wrote: > Check firewall rules and selinux. It sometimes is a pain in the ... :) > 25 lut 2015 01:46 "Barclay Jameson" napisał(a): > > I have tried to install ceph using ceph-deploy but sgdisk seems to >> have too many issues so I did a manual install. After mkfs.btrfs on >> the disks and journals and mounted them I then tried to start the osds >> which failed. The first error was: >> #/etc/init.d/ceph start osd.0 >> /etc/init.d/ceph: osd.0 not found (/etc/ceph/ceph.conf defines , >> /var/lib/ceph defines ) >> >> I then manually added the osds to the conf file with the following as >> an example: >> [osd.0] >> osd_host = node01 >> >> Now when I run the command : >> # /etc/init.d/ceph start osd.0 >> >> There is no error or output from the command and in fact when I do a >> ceph -s no osds are listed as being up. >> Doing as ps aux | grep -i ceph or ps aux | grep -i osd shows there are >> no osd running. >> I also have done htop to see if any process are running and none are >> shown. >> >> I had this working on SL6.5 with Firefly but Giant on Centos 7 has >> been nothing but a giant pain. >> ___ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > > ___ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Centos 7 OSD silently fail to start
I'm having a similar issue. I'm following http://ceph.com/docs/master/install/manual-deployment/ to a T. I have OSDs on the same host deployed with the short-form and they work fine. I am trying to deploy some more via the long form (because I want them to appear in a different location in the crush map). Everything through step 10 (i.e. ceph osd crush add {id-or-name} {weight} [{bucket-type}={bucket-name} ...] ) works just fine. When I go to step 11 (sudo /etc/init.d/ceph start osd.{osd-num}) I get: /etc/init.d/ceph: osd.16 not found (/etc/ceph/ceph.conf defines mon.hobbit01 osd.7 osd.15 osd.10 osd.9 osd.1 osd.14 osd.2 osd.3 osd.13 osd.8 osd.12 osd.6 osd.11 osd.5 osd.4 osd.0 , /var/lib/ceph defines mon.hobbit01 osd.7 osd.15 osd.10 osd.9 osd.1 osd.14 osd.2 osd.3 osd.13 osd.8 osd.12 osd.6 osd.11 osd.5 osd.4 osd.0) On Wed, Feb 25, 2015 at 11:55 AM, Travis Rhoden wrote: > Also, did you successfully start your monitor(s), and define/create the > OSDs within the Ceph cluster itself? > > There are several steps to creating a Ceph cluster manually. I'm unsure > if you have done the steps to actually create and register the OSDs with > the cluster. > > - Travis > > On Wed, Feb 25, 2015 at 9:49 AM, Leszek Master wrote: > >> Check firewall rules and selinux. It sometimes is a pain in the ... :) >> 25 lut 2015 01:46 "Barclay Jameson" napisał(a): >> >> I have tried to install ceph using ceph-deploy but sgdisk seems to >>> have too many issues so I did a manual install. After mkfs.btrfs on >>> the disks and journals and mounted them I then tried to start the osds >>> which failed. The first error was: >>> #/etc/init.d/ceph start osd.0 >>> /etc/init.d/ceph: osd.0 not found (/etc/ceph/ceph.conf defines , >>> /var/lib/ceph defines ) >>> >>> I then manually added the osds to the conf file with the following as >>> an example: >>> [osd.0] >>> osd_host = node01 >>> >>> Now when I run the command : >>> # /etc/init.d/ceph start osd.0 >>> >>> There is no error or output from the command and in fact when I do a >>> ceph -s no osds are listed as being up. >>> Doing as ps aux | grep -i ceph or ps aux | grep -i osd shows there are >>> no osd running. >>> I also have done htop to see if any process are running and none are >>> shown. >>> >>> I had this working on SL6.5 with Firefly but Giant on Centos 7 has >>> been nothing but a giant pain. >>> ___ >>> ceph-users mailing list >>> ceph-users@lists.ceph.com >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>> >> >> ___ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> > > ___ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Centos 7 OSD silently fail to start
I think that your problem lies with systemd (even though you are using SysV syntax, systemd is really doing the work). Systemd does not like multiple arguments and I think this is why it is failing. There is supposed to be some work done to get systemd working ok, but I think it has the limitation of only working with a cluster named 'ceph' currently. What I did to get around the problem was to run the osd command manually: ceph-osd -i Once I understand the under-the-hood stuff, I moved to ceph-disk and now because of the GPT partition IDs, udev automatically starts up the OSD process at boot/creation and moves to the appropiate CRUSH location (configuratble in ceph.conf http://ceph.com/docs/master/rados/operations/crush-map/#crush-location, an example: crush location = host=test rack=rack3 row=row8 datacenter=local region=na-west root=default). To restart an OSD process, I just kill the PID for the OSD then issue ceph-disk activate /dev/sdx1 to restart the OSD process. You probably could stop it with systemctl since I believe udev creates a resource for it (I should probably look into that now that this system will be going production soon). On Wed, Feb 25, 2015 at 2:13 PM, Kyle Hutson wrote: > I'm having a similar issue. > > I'm following http://ceph.com/docs/master/install/manual-deployment/ to a T. > > I have OSDs on the same host deployed with the short-form and they work > fine. I am trying to deploy some more via the long form (because I want them > to appear in a different location in the crush map). Everything through step > 10 (i.e. ceph osd crush add {id-or-name} {weight} > [{bucket-type}={bucket-name} ...] ) works just fine. When I go to step 11 > (sudo /etc/init.d/ceph start osd.{osd-num}) I get: > /etc/init.d/ceph: osd.16 not found (/etc/ceph/ceph.conf defines mon.hobbit01 > osd.7 osd.15 osd.10 osd.9 osd.1 osd.14 osd.2 osd.3 osd.13 osd.8 osd.12 osd.6 > osd.11 osd.5 osd.4 osd.0 , /var/lib/ceph defines mon.hobbit01 osd.7 osd.15 > osd.10 osd.9 osd.1 osd.14 osd.2 osd.3 osd.13 osd.8 osd.12 osd.6 osd.11 osd.5 > osd.4 osd.0) > > > > On Wed, Feb 25, 2015 at 11:55 AM, Travis Rhoden wrote: >> >> Also, did you successfully start your monitor(s), and define/create the >> OSDs within the Ceph cluster itself? >> >> There are several steps to creating a Ceph cluster manually. I'm unsure >> if you have done the steps to actually create and register the OSDs with the >> cluster. >> >> - Travis >> >> On Wed, Feb 25, 2015 at 9:49 AM, Leszek Master wrote: >>> >>> Check firewall rules and selinux. It sometimes is a pain in the ... :) >>> >>> 25 lut 2015 01:46 "Barclay Jameson" napisał(a): >>> I have tried to install ceph using ceph-deploy but sgdisk seems to have too many issues so I did a manual install. After mkfs.btrfs on the disks and journals and mounted them I then tried to start the osds which failed. The first error was: #/etc/init.d/ceph start osd.0 /etc/init.d/ceph: osd.0 not found (/etc/ceph/ceph.conf defines , /var/lib/ceph defines ) I then manually added the osds to the conf file with the following as an example: [osd.0] osd_host = node01 Now when I run the command : # /etc/init.d/ceph start osd.0 There is no error or output from the command and in fact when I do a ceph -s no osds are listed as being up. Doing as ps aux | grep -i ceph or ps aux | grep -i osd shows there are no osd running. I also have done htop to see if any process are running and none are shown. I had this working on SL6.5 with Firefly but Giant on Centos 7 has been nothing but a giant pain. ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>> >>> >>> ___ >>> ceph-users mailing list >>> ceph-users@lists.ceph.com >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>> >> >> >> ___ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > > > ___ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Centos 7 OSD silently fail to start
But I already issued that command (back in step 6). The interesting part is that "ceph-disk activate" apparently does it correctly. Even after reboot, the services start as they should. On Wed, Feb 25, 2015 at 3:54 PM, Robert LeBlanc wrote: > I think that your problem lies with systemd (even though you are using > SysV syntax, systemd is really doing the work). Systemd does not like > multiple arguments and I think this is why it is failing. There is > supposed to be some work done to get systemd working ok, but I think > it has the limitation of only working with a cluster named 'ceph' > currently. > > What I did to get around the problem was to run the osd command manually: > > ceph-osd -i > > Once I understand the under-the-hood stuff, I moved to ceph-disk and > now because of the GPT partition IDs, udev automatically starts up the > OSD process at boot/creation and moves to the appropiate CRUSH > location (configuratble in ceph.conf > http://ceph.com/docs/master/rados/operations/crush-map/#crush-location, > an example: crush location = host=test rack=rack3 row=row8 > datacenter=local region=na-west root=default). To restart an OSD > process, I just kill the PID for the OSD then issue ceph-disk activate > /dev/sdx1 to restart the OSD process. You probably could stop it with > systemctl since I believe udev creates a resource for it (I should > probably look into that now that this system will be going production > soon). > > On Wed, Feb 25, 2015 at 2:13 PM, Kyle Hutson wrote: > > I'm having a similar issue. > > > > I'm following http://ceph.com/docs/master/install/manual-deployment/ to > a T. > > > > I have OSDs on the same host deployed with the short-form and they work > > fine. I am trying to deploy some more via the long form (because I want > them > > to appear in a different location in the crush map). Everything through > step > > 10 (i.e. ceph osd crush add {id-or-name} {weight} > > [{bucket-type}={bucket-name} ...] ) works just fine. When I go to step 11 > > (sudo /etc/init.d/ceph start osd.{osd-num}) I get: > > /etc/init.d/ceph: osd.16 not found (/etc/ceph/ceph.conf defines > mon.hobbit01 > > osd.7 osd.15 osd.10 osd.9 osd.1 osd.14 osd.2 osd.3 osd.13 osd.8 osd.12 > osd.6 > > osd.11 osd.5 osd.4 osd.0 , /var/lib/ceph defines mon.hobbit01 osd.7 > osd.15 > > osd.10 osd.9 osd.1 osd.14 osd.2 osd.3 osd.13 osd.8 osd.12 osd.6 osd.11 > osd.5 > > osd.4 osd.0) > > > > > > > > On Wed, Feb 25, 2015 at 11:55 AM, Travis Rhoden > wrote: > >> > >> Also, did you successfully start your monitor(s), and define/create the > >> OSDs within the Ceph cluster itself? > >> > >> There are several steps to creating a Ceph cluster manually. I'm unsure > >> if you have done the steps to actually create and register the OSDs > with the > >> cluster. > >> > >> - Travis > >> > >> On Wed, Feb 25, 2015 at 9:49 AM, Leszek Master > wrote: > >>> > >>> Check firewall rules and selinux. It sometimes is a pain in the ... :) > >>> > >>> 25 lut 2015 01:46 "Barclay Jameson" > napisał(a): > >>> > I have tried to install ceph using ceph-deploy but sgdisk seems to > have too many issues so I did a manual install. After mkfs.btrfs on > the disks and journals and mounted them I then tried to start the osds > which failed. The first error was: > #/etc/init.d/ceph start osd.0 > /etc/init.d/ceph: osd.0 not found (/etc/ceph/ceph.conf defines , > /var/lib/ceph defines ) > > I then manually added the osds to the conf file with the following as > an example: > [osd.0] > osd_host = node01 > > Now when I run the command : > # /etc/init.d/ceph start osd.0 > > There is no error or output from the command and in fact when I do a > ceph -s no osds are listed as being up. > Doing as ps aux | grep -i ceph or ps aux | grep -i osd shows there are > no osd running. > I also have done htop to see if any process are running and none are > shown. > > I had this working on SL6.5 with Firefly but Giant on Centos 7 has > been nothing but a giant pain. > ___ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >>> > >>> > >>> ___ > >>> ceph-users mailing list > >>> ceph-users@lists.ceph.com > >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >>> > >> > >> > >> ___ > >> ceph-users mailing list > >> ceph-users@lists.ceph.com > >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >> > > > > > > ___ > > ceph-users mailing list > > ceph-users@lists.ceph.com > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph
Re: [ceph-users] Centos 7 OSD silently fail to start
Step #6 in http://ceph.com/docs/master/install/manual-deployment/#long-form only set-ups the file structure for the OSD, it doesn't start the long running process. On Wed, Feb 25, 2015 at 2:59 PM, Kyle Hutson wrote: > But I already issued that command (back in step 6). > > The interesting part is that "ceph-disk activate" apparently does it > correctly. Even after reboot, the services start as they should. > > On Wed, Feb 25, 2015 at 3:54 PM, Robert LeBlanc > wrote: >> >> I think that your problem lies with systemd (even though you are using >> SysV syntax, systemd is really doing the work). Systemd does not like >> multiple arguments and I think this is why it is failing. There is >> supposed to be some work done to get systemd working ok, but I think >> it has the limitation of only working with a cluster named 'ceph' >> currently. >> >> What I did to get around the problem was to run the osd command manually: >> >> ceph-osd -i >> >> Once I understand the under-the-hood stuff, I moved to ceph-disk and >> now because of the GPT partition IDs, udev automatically starts up the >> OSD process at boot/creation and moves to the appropiate CRUSH >> location (configuratble in ceph.conf >> http://ceph.com/docs/master/rados/operations/crush-map/#crush-location, >> an example: crush location = host=test rack=rack3 row=row8 >> datacenter=local region=na-west root=default). To restart an OSD >> process, I just kill the PID for the OSD then issue ceph-disk activate >> /dev/sdx1 to restart the OSD process. You probably could stop it with >> systemctl since I believe udev creates a resource for it (I should >> probably look into that now that this system will be going production >> soon). >> >> On Wed, Feb 25, 2015 at 2:13 PM, Kyle Hutson wrote: >> > I'm having a similar issue. >> > >> > I'm following http://ceph.com/docs/master/install/manual-deployment/ to >> > a T. >> > >> > I have OSDs on the same host deployed with the short-form and they work >> > fine. I am trying to deploy some more via the long form (because I want >> > them >> > to appear in a different location in the crush map). Everything through >> > step >> > 10 (i.e. ceph osd crush add {id-or-name} {weight} >> > [{bucket-type}={bucket-name} ...] ) works just fine. When I go to step >> > 11 >> > (sudo /etc/init.d/ceph start osd.{osd-num}) I get: >> > /etc/init.d/ceph: osd.16 not found (/etc/ceph/ceph.conf defines >> > mon.hobbit01 >> > osd.7 osd.15 osd.10 osd.9 osd.1 osd.14 osd.2 osd.3 osd.13 osd.8 osd.12 >> > osd.6 >> > osd.11 osd.5 osd.4 osd.0 , /var/lib/ceph defines mon.hobbit01 osd.7 >> > osd.15 >> > osd.10 osd.9 osd.1 osd.14 osd.2 osd.3 osd.13 osd.8 osd.12 osd.6 osd.11 >> > osd.5 >> > osd.4 osd.0) >> > >> > >> > >> > On Wed, Feb 25, 2015 at 11:55 AM, Travis Rhoden >> > wrote: >> >> >> >> Also, did you successfully start your monitor(s), and define/create the >> >> OSDs within the Ceph cluster itself? >> >> >> >> There are several steps to creating a Ceph cluster manually. I'm >> >> unsure >> >> if you have done the steps to actually create and register the OSDs >> >> with the >> >> cluster. >> >> >> >> - Travis >> >> >> >> On Wed, Feb 25, 2015 at 9:49 AM, Leszek Master >> >> wrote: >> >>> >> >>> Check firewall rules and selinux. It sometimes is a pain in the ... :) >> >>> >> >>> 25 lut 2015 01:46 "Barclay Jameson" >> >>> napisał(a): >> >>> >> I have tried to install ceph using ceph-deploy but sgdisk seems to >> have too many issues so I did a manual install. After mkfs.btrfs on >> the disks and journals and mounted them I then tried to start the >> osds >> which failed. The first error was: >> #/etc/init.d/ceph start osd.0 >> /etc/init.d/ceph: osd.0 not found (/etc/ceph/ceph.conf defines , >> /var/lib/ceph defines ) >> >> I then manually added the osds to the conf file with the following as >> an example: >> [osd.0] >> osd_host = node01 >> >> Now when I run the command : >> # /etc/init.d/ceph start osd.0 >> >> There is no error or output from the command and in fact when I do a >> ceph -s no osds are listed as being up. >> Doing as ps aux | grep -i ceph or ps aux | grep -i osd shows there >> are >> no osd running. >> I also have done htop to see if any process are running and none are >> shown. >> >> I had this working on SL6.5 with Firefly but Giant on Centos 7 has >> been nothing but a giant pain. >> ___ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >>> >> >>> >> >>> ___ >> >>> ceph-users mailing list >> >>> ceph-users@lists.ceph.com >> >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >>> >> >> >> >> >> >> ___ >> >> ceph-users mailing list >> >> ceph-users@lists.ceph.com
Re: [ceph-users] Centos 7 OSD silently fail to start
So I issue it twice? e.g. ceph-osd -i X --mkfs --mkkey ...other commands... ceph-osd -i X ? On Wed, Feb 25, 2015 at 4:03 PM, Robert LeBlanc wrote: > Step #6 in > http://ceph.com/docs/master/install/manual-deployment/#long-form > only set-ups the file structure for the OSD, it doesn't start the long > running process. > > On Wed, Feb 25, 2015 at 2:59 PM, Kyle Hutson wrote: > > But I already issued that command (back in step 6). > > > > The interesting part is that "ceph-disk activate" apparently does it > > correctly. Even after reboot, the services start as they should. > > > > On Wed, Feb 25, 2015 at 3:54 PM, Robert LeBlanc > > wrote: > >> > >> I think that your problem lies with systemd (even though you are using > >> SysV syntax, systemd is really doing the work). Systemd does not like > >> multiple arguments and I think this is why it is failing. There is > >> supposed to be some work done to get systemd working ok, but I think > >> it has the limitation of only working with a cluster named 'ceph' > >> currently. > >> > >> What I did to get around the problem was to run the osd command > manually: > >> > >> ceph-osd -i > >> > >> Once I understand the under-the-hood stuff, I moved to ceph-disk and > >> now because of the GPT partition IDs, udev automatically starts up the > >> OSD process at boot/creation and moves to the appropiate CRUSH > >> location (configuratble in ceph.conf > >> http://ceph.com/docs/master/rados/operations/crush-map/#crush-location, > >> an example: crush location = host=test rack=rack3 row=row8 > >> datacenter=local region=na-west root=default). To restart an OSD > >> process, I just kill the PID for the OSD then issue ceph-disk activate > >> /dev/sdx1 to restart the OSD process. You probably could stop it with > >> systemctl since I believe udev creates a resource for it (I should > >> probably look into that now that this system will be going production > >> soon). > >> > >> On Wed, Feb 25, 2015 at 2:13 PM, Kyle Hutson > wrote: > >> > I'm having a similar issue. > >> > > >> > I'm following http://ceph.com/docs/master/install/manual-deployment/ > to > >> > a T. > >> > > >> > I have OSDs on the same host deployed with the short-form and they > work > >> > fine. I am trying to deploy some more via the long form (because I > want > >> > them > >> > to appear in a different location in the crush map). Everything > through > >> > step > >> > 10 (i.e. ceph osd crush add {id-or-name} {weight} > >> > [{bucket-type}={bucket-name} ...] ) works just fine. When I go to step > >> > 11 > >> > (sudo /etc/init.d/ceph start osd.{osd-num}) I get: > >> > /etc/init.d/ceph: osd.16 not found (/etc/ceph/ceph.conf defines > >> > mon.hobbit01 > >> > osd.7 osd.15 osd.10 osd.9 osd.1 osd.14 osd.2 osd.3 osd.13 osd.8 osd.12 > >> > osd.6 > >> > osd.11 osd.5 osd.4 osd.0 , /var/lib/ceph defines mon.hobbit01 osd.7 > >> > osd.15 > >> > osd.10 osd.9 osd.1 osd.14 osd.2 osd.3 osd.13 osd.8 osd.12 osd.6 osd.11 > >> > osd.5 > >> > osd.4 osd.0) > >> > > >> > > >> > > >> > On Wed, Feb 25, 2015 at 11:55 AM, Travis Rhoden > >> > wrote: > >> >> > >> >> Also, did you successfully start your monitor(s), and define/create > the > >> >> OSDs within the Ceph cluster itself? > >> >> > >> >> There are several steps to creating a Ceph cluster manually. I'm > >> >> unsure > >> >> if you have done the steps to actually create and register the OSDs > >> >> with the > >> >> cluster. > >> >> > >> >> - Travis > >> >> > >> >> On Wed, Feb 25, 2015 at 9:49 AM, Leszek Master > >> >> wrote: > >> >>> > >> >>> Check firewall rules and selinux. It sometimes is a pain in the ... > :) > >> >>> > >> >>> 25 lut 2015 01:46 "Barclay Jameson" > >> >>> napisał(a): > >> >>> > >> I have tried to install ceph using ceph-deploy but sgdisk seems to > >> have too many issues so I did a manual install. After mkfs.btrfs on > >> the disks and journals and mounted them I then tried to start the > >> osds > >> which failed. The first error was: > >> #/etc/init.d/ceph start osd.0 > >> /etc/init.d/ceph: osd.0 not found (/etc/ceph/ceph.conf defines , > >> /var/lib/ceph defines ) > >> > >> I then manually added the osds to the conf file with the following > as > >> an example: > >> [osd.0] > >> osd_host = node01 > >> > >> Now when I run the command : > >> # /etc/init.d/ceph start osd.0 > >> > >> There is no error or output from the command and in fact when I do > a > >> ceph -s no osds are listed as being up. > >> Doing as ps aux | grep -i ceph or ps aux | grep -i osd shows there > >> are > >> no osd running. > >> I also have done htop to see if any process are running and none > are > >> shown. > >> > >> I had this working on SL6.5 with Firefly but Giant on Centos 7 has > >> been nothing but a giant pain. > >> ___ > >> ceph-users mailing list > >> ceph-users@
Re: [ceph-users] Centos 7 OSD silently fail to start
I just tried it, and that does indeed get the OSD to start. However, it doesn't add it to the appropriate place so it would survive a reboot. In my case, running 'service ceph status osd.16' still results in the same line I posted above. There's still something broken such that 'ceph-disk activate' works correctly, but using the long-form version does not. On Wed, Feb 25, 2015 at 4:03 PM, Robert LeBlanc wrote: > Step #6 in > http://ceph.com/docs/master/install/manual-deployment/#long-form > only set-ups the file structure for the OSD, it doesn't start the long > running process. > > On Wed, Feb 25, 2015 at 2:59 PM, Kyle Hutson wrote: > > But I already issued that command (back in step 6). > > > > The interesting part is that "ceph-disk activate" apparently does it > > correctly. Even after reboot, the services start as they should. > > > > On Wed, Feb 25, 2015 at 3:54 PM, Robert LeBlanc > > wrote: > >> > >> I think that your problem lies with systemd (even though you are using > >> SysV syntax, systemd is really doing the work). Systemd does not like > >> multiple arguments and I think this is why it is failing. There is > >> supposed to be some work done to get systemd working ok, but I think > >> it has the limitation of only working with a cluster named 'ceph' > >> currently. > >> > >> What I did to get around the problem was to run the osd command > manually: > >> > >> ceph-osd -i > >> > >> Once I understand the under-the-hood stuff, I moved to ceph-disk and > >> now because of the GPT partition IDs, udev automatically starts up the > >> OSD process at boot/creation and moves to the appropiate CRUSH > >> location (configuratble in ceph.conf > >> http://ceph.com/docs/master/rados/operations/crush-map/#crush-location, > >> an example: crush location = host=test rack=rack3 row=row8 > >> datacenter=local region=na-west root=default). To restart an OSD > >> process, I just kill the PID for the OSD then issue ceph-disk activate > >> /dev/sdx1 to restart the OSD process. You probably could stop it with > >> systemctl since I believe udev creates a resource for it (I should > >> probably look into that now that this system will be going production > >> soon). > >> > >> On Wed, Feb 25, 2015 at 2:13 PM, Kyle Hutson > wrote: > >> > I'm having a similar issue. > >> > > >> > I'm following http://ceph.com/docs/master/install/manual-deployment/ > to > >> > a T. > >> > > >> > I have OSDs on the same host deployed with the short-form and they > work > >> > fine. I am trying to deploy some more via the long form (because I > want > >> > them > >> > to appear in a different location in the crush map). Everything > through > >> > step > >> > 10 (i.e. ceph osd crush add {id-or-name} {weight} > >> > [{bucket-type}={bucket-name} ...] ) works just fine. When I go to step > >> > 11 > >> > (sudo /etc/init.d/ceph start osd.{osd-num}) I get: > >> > /etc/init.d/ceph: osd.16 not found (/etc/ceph/ceph.conf defines > >> > mon.hobbit01 > >> > osd.7 osd.15 osd.10 osd.9 osd.1 osd.14 osd.2 osd.3 osd.13 osd.8 osd.12 > >> > osd.6 > >> > osd.11 osd.5 osd.4 osd.0 , /var/lib/ceph defines mon.hobbit01 osd.7 > >> > osd.15 > >> > osd.10 osd.9 osd.1 osd.14 osd.2 osd.3 osd.13 osd.8 osd.12 osd.6 osd.11 > >> > osd.5 > >> > osd.4 osd.0) > >> > > >> > > >> > > >> > On Wed, Feb 25, 2015 at 11:55 AM, Travis Rhoden > >> > wrote: > >> >> > >> >> Also, did you successfully start your monitor(s), and define/create > the > >> >> OSDs within the Ceph cluster itself? > >> >> > >> >> There are several steps to creating a Ceph cluster manually. I'm > >> >> unsure > >> >> if you have done the steps to actually create and register the OSDs > >> >> with the > >> >> cluster. > >> >> > >> >> - Travis > >> >> > >> >> On Wed, Feb 25, 2015 at 9:49 AM, Leszek Master > >> >> wrote: > >> >>> > >> >>> Check firewall rules and selinux. It sometimes is a pain in the ... > :) > >> >>> > >> >>> 25 lut 2015 01:46 "Barclay Jameson" > >> >>> napisał(a): > >> >>> > >> I have tried to install ceph using ceph-deploy but sgdisk seems to > >> have too many issues so I did a manual install. After mkfs.btrfs on > >> the disks and journals and mounted them I then tried to start the > >> osds > >> which failed. The first error was: > >> #/etc/init.d/ceph start osd.0 > >> /etc/init.d/ceph: osd.0 not found (/etc/ceph/ceph.conf defines , > >> /var/lib/ceph defines ) > >> > >> I then manually added the osds to the conf file with the following > as > >> an example: > >> [osd.0] > >> osd_host = node01 > >> > >> Now when I run the command : > >> # /etc/init.d/ceph start osd.0 > >> > >> There is no error or output from the command and in fact when I do > a > >> ceph -s no osds are listed as being up. > >> Doing as ps aux | grep -i ceph or ps aux | grep -i osd shows there > >> are > >> no osd running. > >> I also have done htop to see if any process are running and no
Re: [ceph-users] Centos 7 OSD silently fail to start
I am using the long form and have it working. The one thing that I saw was to change from osd_host to just host. See if that works. On Feb 25, 2015 5:44 PM, "Kyle Hutson" wrote: > I just tried it, and that does indeed get the OSD to start. > > However, it doesn't add it to the appropriate place so it would survive a > reboot. In my case, running 'service ceph status osd.16' still results in > the same line I posted above. > > There's still something broken such that 'ceph-disk activate' works > correctly, but using the long-form version does not. > > On Wed, Feb 25, 2015 at 4:03 PM, Robert LeBlanc > wrote: > >> Step #6 in >> http://ceph.com/docs/master/install/manual-deployment/#long-form >> only set-ups the file structure for the OSD, it doesn't start the long >> running process. >> >> On Wed, Feb 25, 2015 at 2:59 PM, Kyle Hutson wrote: >> > But I already issued that command (back in step 6). >> > >> > The interesting part is that "ceph-disk activate" apparently does it >> > correctly. Even after reboot, the services start as they should. >> > >> > On Wed, Feb 25, 2015 at 3:54 PM, Robert LeBlanc >> > wrote: >> >> >> >> I think that your problem lies with systemd (even though you are using >> >> SysV syntax, systemd is really doing the work). Systemd does not like >> >> multiple arguments and I think this is why it is failing. There is >> >> supposed to be some work done to get systemd working ok, but I think >> >> it has the limitation of only working with a cluster named 'ceph' >> >> currently. >> >> >> >> What I did to get around the problem was to run the osd command >> manually: >> >> >> >> ceph-osd -i >> >> >> >> Once I understand the under-the-hood stuff, I moved to ceph-disk and >> >> now because of the GPT partition IDs, udev automatically starts up the >> >> OSD process at boot/creation and moves to the appropiate CRUSH >> >> location (configuratble in ceph.conf >> >> http://ceph.com/docs/master/rados/operations/crush-map/#crush-location >> , >> >> an example: crush location = host=test rack=rack3 row=row8 >> >> datacenter=local region=na-west root=default). To restart an OSD >> >> process, I just kill the PID for the OSD then issue ceph-disk activate >> >> /dev/sdx1 to restart the OSD process. You probably could stop it with >> >> systemctl since I believe udev creates a resource for it (I should >> >> probably look into that now that this system will be going production >> >> soon). >> >> >> >> On Wed, Feb 25, 2015 at 2:13 PM, Kyle Hutson >> wrote: >> >> > I'm having a similar issue. >> >> > >> >> > I'm following http://ceph.com/docs/master/install/manual-deployment/ >> to >> >> > a T. >> >> > >> >> > I have OSDs on the same host deployed with the short-form and they >> work >> >> > fine. I am trying to deploy some more via the long form (because I >> want >> >> > them >> >> > to appear in a different location in the crush map). Everything >> through >> >> > step >> >> > 10 (i.e. ceph osd crush add {id-or-name} {weight} >> >> > [{bucket-type}={bucket-name} ...] ) works just fine. When I go to >> step >> >> > 11 >> >> > (sudo /etc/init.d/ceph start osd.{osd-num}) I get: >> >> > /etc/init.d/ceph: osd.16 not found (/etc/ceph/ceph.conf defines >> >> > mon.hobbit01 >> >> > osd.7 osd.15 osd.10 osd.9 osd.1 osd.14 osd.2 osd.3 osd.13 osd.8 >> osd.12 >> >> > osd.6 >> >> > osd.11 osd.5 osd.4 osd.0 , /var/lib/ceph defines mon.hobbit01 osd.7 >> >> > osd.15 >> >> > osd.10 osd.9 osd.1 osd.14 osd.2 osd.3 osd.13 osd.8 osd.12 osd.6 >> osd.11 >> >> > osd.5 >> >> > osd.4 osd.0) >> >> > >> >> > >> >> > >> >> > On Wed, Feb 25, 2015 at 11:55 AM, Travis Rhoden >> >> > wrote: >> >> >> >> >> >> Also, did you successfully start your monitor(s), and define/create >> the >> >> >> OSDs within the Ceph cluster itself? >> >> >> >> >> >> There are several steps to creating a Ceph cluster manually. I'm >> >> >> unsure >> >> >> if you have done the steps to actually create and register the OSDs >> >> >> with the >> >> >> cluster. >> >> >> >> >> >> - Travis >> >> >> >> >> >> On Wed, Feb 25, 2015 at 9:49 AM, Leszek Master >> >> >> wrote: >> >> >>> >> >> >>> Check firewall rules and selinux. It sometimes is a pain in the >> ... :) >> >> >>> >> >> >>> 25 lut 2015 01:46 "Barclay Jameson" >> >> >>> napisał(a): >> >> >>> >> >> I have tried to install ceph using ceph-deploy but sgdisk seems to >> >> have too many issues so I did a manual install. After mkfs.btrfs >> on >> >> the disks and journals and mounted them I then tried to start the >> >> osds >> >> which failed. The first error was: >> >> #/etc/init.d/ceph start osd.0 >> >> /etc/init.d/ceph: osd.0 not found (/etc/ceph/ceph.conf defines , >> >> /var/lib/ceph defines ) >> >> >> >> I then manually added the osds to the conf file with the >> following as >> >> an example: >> >> [osd.0] >> >> osd_host = node01 >> >> >> >> Now when I run the command : >> >> # /etc/init.d/ceph start osd.0 >
Re: [ceph-users] Centos 7 OSD silently fail to start
Heres the doc I used to get the info: http://ceph.com/docs/master/rados/troubleshooting/troubleshooting-osd/ On Feb 25, 2015 5:55 PM, "Thomas Foster" wrote: > I am using the long form and have it working. The one thing that I saw > was to change from osd_host to just host. See if that works. > On Feb 25, 2015 5:44 PM, "Kyle Hutson" wrote: > >> I just tried it, and that does indeed get the OSD to start. >> >> However, it doesn't add it to the appropriate place so it would survive a >> reboot. In my case, running 'service ceph status osd.16' still results in >> the same line I posted above. >> >> There's still something broken such that 'ceph-disk activate' works >> correctly, but using the long-form version does not. >> >> On Wed, Feb 25, 2015 at 4:03 PM, Robert LeBlanc >> wrote: >> >>> Step #6 in >>> http://ceph.com/docs/master/install/manual-deployment/#long-form >>> only set-ups the file structure for the OSD, it doesn't start the long >>> running process. >>> >>> On Wed, Feb 25, 2015 at 2:59 PM, Kyle Hutson wrote: >>> > But I already issued that command (back in step 6). >>> > >>> > The interesting part is that "ceph-disk activate" apparently does it >>> > correctly. Even after reboot, the services start as they should. >>> > >>> > On Wed, Feb 25, 2015 at 3:54 PM, Robert LeBlanc >>> > wrote: >>> >> >>> >> I think that your problem lies with systemd (even though you are using >>> >> SysV syntax, systemd is really doing the work). Systemd does not like >>> >> multiple arguments and I think this is why it is failing. There is >>> >> supposed to be some work done to get systemd working ok, but I think >>> >> it has the limitation of only working with a cluster named 'ceph' >>> >> currently. >>> >> >>> >> What I did to get around the problem was to run the osd command >>> manually: >>> >> >>> >> ceph-osd -i >>> >> >>> >> Once I understand the under-the-hood stuff, I moved to ceph-disk and >>> >> now because of the GPT partition IDs, udev automatically starts up the >>> >> OSD process at boot/creation and moves to the appropiate CRUSH >>> >> location (configuratble in ceph.conf >>> >> >>> http://ceph.com/docs/master/rados/operations/crush-map/#crush-location, >>> >> an example: crush location = host=test rack=rack3 row=row8 >>> >> datacenter=local region=na-west root=default). To restart an OSD >>> >> process, I just kill the PID for the OSD then issue ceph-disk activate >>> >> /dev/sdx1 to restart the OSD process. You probably could stop it with >>> >> systemctl since I believe udev creates a resource for it (I should >>> >> probably look into that now that this system will be going production >>> >> soon). >>> >> >>> >> On Wed, Feb 25, 2015 at 2:13 PM, Kyle Hutson >>> wrote: >>> >> > I'm having a similar issue. >>> >> > >>> >> > I'm following >>> http://ceph.com/docs/master/install/manual-deployment/ to >>> >> > a T. >>> >> > >>> >> > I have OSDs on the same host deployed with the short-form and they >>> work >>> >> > fine. I am trying to deploy some more via the long form (because I >>> want >>> >> > them >>> >> > to appear in a different location in the crush map). Everything >>> through >>> >> > step >>> >> > 10 (i.e. ceph osd crush add {id-or-name} {weight} >>> >> > [{bucket-type}={bucket-name} ...] ) works just fine. When I go to >>> step >>> >> > 11 >>> >> > (sudo /etc/init.d/ceph start osd.{osd-num}) I get: >>> >> > /etc/init.d/ceph: osd.16 not found (/etc/ceph/ceph.conf defines >>> >> > mon.hobbit01 >>> >> > osd.7 osd.15 osd.10 osd.9 osd.1 osd.14 osd.2 osd.3 osd.13 osd.8 >>> osd.12 >>> >> > osd.6 >>> >> > osd.11 osd.5 osd.4 osd.0 , /var/lib/ceph defines mon.hobbit01 osd.7 >>> >> > osd.15 >>> >> > osd.10 osd.9 osd.1 osd.14 osd.2 osd.3 osd.13 osd.8 osd.12 osd.6 >>> osd.11 >>> >> > osd.5 >>> >> > osd.4 osd.0) >>> >> > >>> >> > >>> >> > >>> >> > On Wed, Feb 25, 2015 at 11:55 AM, Travis Rhoden >>> >> > wrote: >>> >> >> >>> >> >> Also, did you successfully start your monitor(s), and >>> define/create the >>> >> >> OSDs within the Ceph cluster itself? >>> >> >> >>> >> >> There are several steps to creating a Ceph cluster manually. I'm >>> >> >> unsure >>> >> >> if you have done the steps to actually create and register the OSDs >>> >> >> with the >>> >> >> cluster. >>> >> >> >>> >> >> - Travis >>> >> >> >>> >> >> On Wed, Feb 25, 2015 at 9:49 AM, Leszek Master >>> >> >> wrote: >>> >> >>> >>> >> >>> Check firewall rules and selinux. It sometimes is a pain in the >>> ... :) >>> >> >>> >>> >> >>> 25 lut 2015 01:46 "Barclay Jameson" >>> >> >>> napisał(a): >>> >> >>> >>> >> I have tried to install ceph using ceph-deploy but sgdisk seems >>> to >>> >> have too many issues so I did a manual install. After mkfs.btrfs >>> on >>> >> the disks and journals and mounted them I then tried to start the >>> >> osds >>> >> which failed. The first error was: >>> >> #/etc/init.d/ceph start osd.0 >>> >> /etc/init.d/ceph: osd.0 not found (/etc/ceph/ceph.conf defines , >>> >>
Re: [ceph-users] Centos 7 OSD silently fail to start
Thank you Thomas. You at least made me look it the right spot. Their long-form is showing what to do for a mon, not an osd. At the bottom of step 11, instead of sudo touch /var/lib/ceph/mon/{cluster-name}-{hostname}/sysvinit It should read sudo touch /var/lib/ceph/osd/{cluster-name}-{osd-num}/sysvinit Once I did that 'service ceph status' correctly shows that I have that OSD available and I can start or stop it from there. On Wed, Feb 25, 2015 at 4:55 PM, Thomas Foster wrote: > I am using the long form and have it working. The one thing that I saw > was to change from osd_host to just host. See if that works. > On Feb 25, 2015 5:44 PM, "Kyle Hutson" wrote: > >> I just tried it, and that does indeed get the OSD to start. >> >> However, it doesn't add it to the appropriate place so it would survive a >> reboot. In my case, running 'service ceph status osd.16' still results in >> the same line I posted above. >> >> There's still something broken such that 'ceph-disk activate' works >> correctly, but using the long-form version does not. >> >> On Wed, Feb 25, 2015 at 4:03 PM, Robert LeBlanc >> wrote: >> >>> Step #6 in >>> http://ceph.com/docs/master/install/manual-deployment/#long-form >>> only set-ups the file structure for the OSD, it doesn't start the long >>> running process. >>> >>> On Wed, Feb 25, 2015 at 2:59 PM, Kyle Hutson wrote: >>> > But I already issued that command (back in step 6). >>> > >>> > The interesting part is that "ceph-disk activate" apparently does it >>> > correctly. Even after reboot, the services start as they should. >>> > >>> > On Wed, Feb 25, 2015 at 3:54 PM, Robert LeBlanc >>> > wrote: >>> >> >>> >> I think that your problem lies with systemd (even though you are using >>> >> SysV syntax, systemd is really doing the work). Systemd does not like >>> >> multiple arguments and I think this is why it is failing. There is >>> >> supposed to be some work done to get systemd working ok, but I think >>> >> it has the limitation of only working with a cluster named 'ceph' >>> >> currently. >>> >> >>> >> What I did to get around the problem was to run the osd command >>> manually: >>> >> >>> >> ceph-osd -i >>> >> >>> >> Once I understand the under-the-hood stuff, I moved to ceph-disk and >>> >> now because of the GPT partition IDs, udev automatically starts up the >>> >> OSD process at boot/creation and moves to the appropiate CRUSH >>> >> location (configuratble in ceph.conf >>> >> >>> http://ceph.com/docs/master/rados/operations/crush-map/#crush-location, >>> >> an example: crush location = host=test rack=rack3 row=row8 >>> >> datacenter=local region=na-west root=default). To restart an OSD >>> >> process, I just kill the PID for the OSD then issue ceph-disk activate >>> >> /dev/sdx1 to restart the OSD process. You probably could stop it with >>> >> systemctl since I believe udev creates a resource for it (I should >>> >> probably look into that now that this system will be going production >>> >> soon). >>> >> >>> >> On Wed, Feb 25, 2015 at 2:13 PM, Kyle Hutson >>> wrote: >>> >> > I'm having a similar issue. >>> >> > >>> >> > I'm following >>> http://ceph.com/docs/master/install/manual-deployment/ to >>> >> > a T. >>> >> > >>> >> > I have OSDs on the same host deployed with the short-form and they >>> work >>> >> > fine. I am trying to deploy some more via the long form (because I >>> want >>> >> > them >>> >> > to appear in a different location in the crush map). Everything >>> through >>> >> > step >>> >> > 10 (i.e. ceph osd crush add {id-or-name} {weight} >>> >> > [{bucket-type}={bucket-name} ...] ) works just fine. When I go to >>> step >>> >> > 11 >>> >> > (sudo /etc/init.d/ceph start osd.{osd-num}) I get: >>> >> > /etc/init.d/ceph: osd.16 not found (/etc/ceph/ceph.conf defines >>> >> > mon.hobbit01 >>> >> > osd.7 osd.15 osd.10 osd.9 osd.1 osd.14 osd.2 osd.3 osd.13 osd.8 >>> osd.12 >>> >> > osd.6 >>> >> > osd.11 osd.5 osd.4 osd.0 , /var/lib/ceph defines mon.hobbit01 osd.7 >>> >> > osd.15 >>> >> > osd.10 osd.9 osd.1 osd.14 osd.2 osd.3 osd.13 osd.8 osd.12 osd.6 >>> osd.11 >>> >> > osd.5 >>> >> > osd.4 osd.0) >>> >> > >>> >> > >>> >> > >>> >> > On Wed, Feb 25, 2015 at 11:55 AM, Travis Rhoden >>> >> > wrote: >>> >> >> >>> >> >> Also, did you successfully start your monitor(s), and >>> define/create the >>> >> >> OSDs within the Ceph cluster itself? >>> >> >> >>> >> >> There are several steps to creating a Ceph cluster manually. I'm >>> >> >> unsure >>> >> >> if you have done the steps to actually create and register the OSDs >>> >> >> with the >>> >> >> cluster. >>> >> >> >>> >> >> - Travis >>> >> >> >>> >> >> On Wed, Feb 25, 2015 at 9:49 AM, Leszek Master >>> >> >> wrote: >>> >> >>> >>> >> >>> Check firewall rules and selinux. It sometimes is a pain in the >>> ... :) >>> >> >>> >>> >> >>> 25 lut 2015 01:46 "Barclay Jameson" >>> >> >>> napisał(a): >>> >> >>> >>> >> I have tried to install ceph using ceph-deploy but sgdisk seems >>> to >>> >> have too many is
Re: [ceph-users] Centos 7 OSD silently fail to start
On 02/26/2015 09:05 AM, Kyle Hutson wrote: Thank you Thomas. You at least made me look it the right spot. Their long-form is showing what to do for a mon, not an osd. At the bottom of step 11, instead of sudo touch /var/lib/ceph/mon/{cluster-name}-{hostname}/sysvinit It should read sudo touch /var/lib/ceph/osd/{cluster-name}-{osd-num}/sysvinit Once I did that 'service ceph status' correctly shows that I have that OSD available and I can start or stop it from there. Could you open a bug for this? Cheers, Brad ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Centos 7 OSD silently fail to start
Just did it. Thanks for suggesting it. On Wed, Feb 25, 2015 at 5:59 PM, Brad Hubbard wrote: > On 02/26/2015 09:05 AM, Kyle Hutson wrote: > >> Thank you Thomas. You at least made me look it the right spot. Their >> long-form is showing what to do for a mon, not an osd. >> >> At the bottom of step 11, instead of >> sudo touch /var/lib/ceph/mon/{cluster-name}-{hostname}/sysvinit >> >> It should read >> sudo touch /var/lib/ceph/osd/{cluster-name}-{osd-num}/sysvinit >> >> Once I did that 'service ceph status' correctly shows that I have that >> OSD available and I can start or stop it from there. >> >> > Could you open a bug for this? > > Cheers, > Brad > > ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Centos 7 OSD silently fail to start
On 02/26/2015 03:24 PM, Kyle Hutson wrote: Just did it. Thanks for suggesting it. No, definitely thank you. Much appreciated. On Wed, Feb 25, 2015 at 5:59 PM, Brad Hubbard mailto:bhubb...@redhat.com>> wrote: On 02/26/2015 09:05 AM, Kyle Hutson wrote: Thank you Thomas. You at least made me look it the right spot. Their long-form is showing what to do for a mon, not an osd. At the bottom of step 11, instead of sudo touch /var/lib/ceph/mon/{cluster-__name}-{hostname}/sysvinit It should read sudo touch /var/lib/ceph/osd/{cluster-__name}-{osd-num}/sysvinit Once I did that 'service ceph status' correctly shows that I have that OSD available and I can start or stop it from there. Could you open a bug for this? Cheers, Brad ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com