RE: Remote Ceph Install

2012-12-03 Thread Blackwell, Edward
Hi Dan,

Thanks for the welcome and the advice.  There indeed was a problem with the 
host name and capitalization as you described, but once I corrected that, a new 
issue began to occur when I ran the ceph-deploy mon command.  The command 
appears to run successfully (no output from the command is generated), but when 
I check the status on one of the servers in the cluster (ceph04 and ELSCEPH01) 
as recommended by the directions, I get the following:

root@cephclient01:~/my-admin-sandbox# ceph-deploy mon
root@cephclient01:~/my-admin-sandbox# ssh ceph04 ceph -s
2012-12-03 13:26:30.031854 7fc353d60780 -1 auth: failed to open keyring from 
/etc/ceph/ceph.client.admin.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin
2012-12-03 13:26:30.031963 7fc353d60780 -1 monclient(hunting): failed to open 
keyring: (2) No such file or directory
2012-12-03 13:26:30.032042 7fc353d60780 -1 ceph_tool_common_init failed.
root@cephclient01:~/my-admin-sandbox#

Behind the scenes, is the ceph-deploy mon command executing the mkcephfs 
command, which creates the keyring file?  If so, could that command be failing 
somehow, and hence the status command is not able to return the status of the 
Ceph installation?

I even tried executing the ceph-deploy mon command using the -v option, and 
got the following, so it seems to be working correctly:

root@cephclient01:~/my-admin-sandbox# ceph-deploy -v mon
DEBUG:ceph_deploy.mon:Deploying mon, cluster ceph hosts ceph04 ELSCEPH01
DEBUG:ceph_deploy.mon:Deploying mon to ceph04
DEBUG:ceph_deploy.mon:Deploying mon to ELSCEPH01
root@cephclient01:~/my-admin-sandbox#

I'm at a loss as to what to check or what do next to get past this situation.  
Any help would be greatly appreciated.

Thanks,

Todd

Todd Blackwell 
HARRIS Corporation 
  321-984-6911 
 954-817-3662 
  eblac...@harris.com 

-Original Message-
From: ceph-devel-ow...@vger.kernel.org 
[mailto:ceph-devel-ow...@vger.kernel.org] On Behalf Of Dan Mick
Sent: Monday, November 19, 2012 10:47 PM
To: Blackwell, Edward
Cc: ceph-devel@vger.kernel.org
Subject: Re: Remote Ceph Install



On 11/19/2012 11:42 AM, Blackwell, Edward wrote:
 Hi,
 I work for Harris Corporation, and we are investigating Ceph as a potential 
 solution to a storage problem that one of our government customers is 
 currently having.  I've already created a two-node cluster on a couple of VMs 
 with another VM acting as an administrative client.  The cluster was created 
 using some installation instructions supplied to us via Inktank, and through 
 the use of the ceph-deploy script.  Aside from a couple of quirky 
 discrepancies between the installation instructions and my environment, 
 everything went well.  My issue has cropped up on the second cluster I'm 
 trying to create, which is using a VM and a non-VM server for the nodes in 
 the cluster.  Eventually, both nodes in this cluster will be non-VMs, but 
 we're still waiting on the hardware for the second node, so I'm using a VM in 
 the meantime just to get this second cluster up and going.  Of course, the 
 administrative client node is still a VM.

Hi Ed.  Welcome.

 The problem that I'm having with this second cluster concerns the non-VM 
 server (elsceph01 for the sake of the commands mentioned from here on out).  
 In particular, the issue crops up with the ceph-deploy install elsceph01 
 command I'm executing on my client VM (cephclient01) to install Ceph on the 
 non-VM server. The installation doesn't appear to be working as the command 
 does not return the OK message that it should when it completes successfully. 
  I've tried using the verbose option on the command to see if that sheds any 
 light on the subject, but alas, it does not:


 root@cephclient01:~/my-admin-sandbox# ceph-deploy -v install elsceph01
 DEBUG:ceph_deploy.install:Installing stable version argonaut on cluster ceph 
 hosts elsceph01
 DEBUG:ceph_deploy.install:Detecting platform for host elsceph01 ...
 DEBUG:ceph_deploy.install:Installing for Ubuntu 12.04 on host elsceph01 ...
 root@cephclient01:~/my-admin-sandbox#


 Would you happen to have a breakdown of the commands being executed by the 
 ceph-deploy script behind the scenes so I can maybe execute them one-by-one 
 to see where the error is?  I have confirmed that it looks like the 
 installation of the software has succeeded as I did a which ceph command on 
 elsceph01, and it reported back /usr/bin/ceph.  Also, /etc/ceph/ceph.conf is 
 there, and it matches the file created by the ceph-deploy new ... command on 
 the client.  Does the install command do a mkcephfs behind the scenes?  The 
 reason I ask is that when I do the ceph-deploy mon command from the client, 
 which is the next command listed in the instructions to do, I get this output:

Basically install just runs the appropriate debian package commands to 
get the requested release of Ceph installed on the target host (in this 
case, defaulting to argonaut).  The command normally doesn't

Re: Remote Ceph Install

2012-12-03 Thread Dan Mick



On 12/03/2012 10:53 AM, Blackwell, Edward wrote:

Hi Dan,

Thanks for the welcome and the advice.  There indeed was a problem with the host name and 
capitalization as you described, but once I corrected that, a new issue began to occur 
when I ran the ceph-deploy mon command.  The command appears to run 
successfully (no output from the command is generated), but when I check the status on 
one of the servers in the cluster (ceph04 and ELSCEPH01) as recommended by the 
directions, I get the following:

root@cephclient01:~/my-admin-sandbox# ceph-deploy mon
root@cephclient01:~/my-admin-sandbox# ssh ceph04 ceph -s
2012-12-03 13:26:30.031854 7fc353d60780 -1 auth: failed to open keyring from 
/etc/ceph/ceph.client.admin.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin
2012-12-03 13:26:30.031963 7fc353d60780 -1 monclient(hunting): failed to open 
keyring: (2) No such file or directory
2012-12-03 13:26:30.032042 7fc353d60780 -1 ceph_tool_common_init failed.
root@cephclient01:~/my-admin-sandbox#


Yeah, looks like the keys didn't get distributed correctly.  Did you do 
the ceph-deploy gatherkeys step?




Behind the scenes, is the ceph-deploy mon command executing the mkcephfs 
command, which creates the keyring file?  If so, could that command be failing somehow, 
and hence the status command is not able to return the status of the Ceph installation?

I even tried executing the ceph-deploy mon command using the -v option, and 
got the following, so it seems to be working correctly:

root@cephclient01:~/my-admin-sandbox# ceph-deploy -v mon
DEBUG:ceph_deploy.mon:Deploying mon, cluster ceph hosts ceph04 ELSCEPH01
DEBUG:ceph_deploy.mon:Deploying mon to ceph04
DEBUG:ceph_deploy.mon:Deploying mon to ELSCEPH01
root@cephclient01:~/my-admin-sandbox#

I'm at a loss as to what to check or what do next to get past this situation.  
Any help would be greatly appreciated.

Thanks,

Todd

Todd Blackwell
HARRIS Corporation
   321-984-6911
 954-817-3662
   eblac...@harris.com

-Original Message-
From: ceph-devel-ow...@vger.kernel.org 
[mailto:ceph-devel-ow...@vger.kernel.org] On Behalf Of Dan Mick
Sent: Monday, November 19, 2012 10:47 PM
To: Blackwell, Edward
Cc: ceph-devel@vger.kernel.org
Subject: Re: Remote Ceph Install



On 11/19/2012 11:42 AM, Blackwell, Edward wrote:

Hi,
I work for Harris Corporation, and we are investigating Ceph as a potential 
solution to a storage problem that one of our government customers is currently 
having.  I've already created a two-node cluster on a couple of VMs with 
another VM acting as an administrative client.  The cluster was created using 
some installation instructions supplied to us via Inktank, and through the use 
of the ceph-deploy script.  Aside from a couple of quirky discrepancies between 
the installation instructions and my environment, everything went well.  My 
issue has cropped up on the second cluster I'm trying to create, which is using 
a VM and a non-VM server for the nodes in the cluster.  Eventually, both nodes 
in this cluster will be non-VMs, but we're still waiting on the hardware for 
the second node, so I'm using a VM in the meantime just to get this second 
cluster up and going.  Of course, the administrative client node is still a VM.


Hi Ed.  Welcome.


The problem that I'm having with this second cluster concerns the non-VM server 
(elsceph01 for the sake of the commands mentioned from here on out).  In 
particular, the issue crops up with the ceph-deploy install elsceph01 command 
I'm executing on my client VM (cephclient01) to install Ceph on the non-VM 
server. The installation doesn't appear to be working as the command does not 
return the OK message that it should when it completes successfully.  I've 
tried using the verbose option on the command to see if that sheds any light on 
the subject, but alas, it does not:


root@cephclient01:~/my-admin-sandbox# ceph-deploy -v install elsceph01
DEBUG:ceph_deploy.install:Installing stable version argonaut on cluster ceph 
hosts elsceph01
DEBUG:ceph_deploy.install:Detecting platform for host elsceph01 ...
DEBUG:ceph_deploy.install:Installing for Ubuntu 12.04 on host elsceph01 ...
root@cephclient01:~/my-admin-sandbox#


Would you happen to have a breakdown of the commands being executed by the 
ceph-deploy script behind the scenes so I can maybe execute them one-by-one to 
see where the error is?  I have confirmed that it looks like the installation 
of the software has succeeded as I did a which ceph command on elsceph01, and 
it reported back /usr/bin/ceph.  Also, /etc/ceph/ceph.conf is there, and it 
matches the file created by the ceph-deploy new ... command on the client.  
Does the install command do a mkcephfs behind the scenes?  The reason I ask is 
that when I do the ceph-deploy mon command from the client, which is the next 
command listed in the instructions to do, I get this output:


Basically install just runs the appropriate debian package commands to
get

RE: Remote Ceph Install

2012-12-03 Thread Blackwell, Edward
Hi Dan,

In the version of the Ceph installation instructions I was given, I don't have 
a ceph-deploy gatherkeys step.  Is there a newer version, or can you briefly 
describe the use of this command?

Thanks,

Todd

Todd Blackwell 
HARRIS Corporation 
  321-984-6911 
 954-817-3662 
  eblac...@harris.com 


-Original Message-
From: Dan Mick [mailto:dan.m...@inktank.com] 
Sent: Monday, December 03, 2012 3:37 PM
To: Blackwell, Edward
Cc: ceph-devel@vger.kernel.org
Subject: Re: Remote Ceph Install



On 12/03/2012 10:53 AM, Blackwell, Edward wrote:
 Hi Dan,

 Thanks for the welcome and the advice.  There indeed was a problem with the 
 host name and capitalization as you described, but once I corrected that, a 
 new issue began to occur when I ran the ceph-deploy mon command.  The 
 command appears to run successfully (no output from the command is 
 generated), but when I check the status on one of the servers in the cluster 
 (ceph04 and ELSCEPH01) as recommended by the directions, I get the following:

 root@cephclient01:~/my-admin-sandbox# ceph-deploy mon
 root@cephclient01:~/my-admin-sandbox# ssh ceph04 ceph -s
 2012-12-03 13:26:30.031854 7fc353d60780 -1 auth: failed to open keyring from 
 /etc/ceph/ceph.client.admin.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin
 2012-12-03 13:26:30.031963 7fc353d60780 -1 monclient(hunting): failed to open 
 keyring: (2) No such file or directory
 2012-12-03 13:26:30.032042 7fc353d60780 -1 ceph_tool_common_init failed.
 root@cephclient01:~/my-admin-sandbox#

Yeah, looks like the keys didn't get distributed correctly.  Did you do 
the ceph-deploy gatherkeys step?


 Behind the scenes, is the ceph-deploy mon command executing the mkcephfs 
 command, which creates the keyring file?  If so, could that command be 
 failing somehow, and hence the status command is not able to return the 
 status of the Ceph installation?

 I even tried executing the ceph-deploy mon command using the -v option, and 
 got the following, so it seems to be working correctly:

 root@cephclient01:~/my-admin-sandbox# ceph-deploy -v mon
 DEBUG:ceph_deploy.mon:Deploying mon, cluster ceph hosts ceph04 ELSCEPH01
 DEBUG:ceph_deploy.mon:Deploying mon to ceph04
 DEBUG:ceph_deploy.mon:Deploying mon to ELSCEPH01
 root@cephclient01:~/my-admin-sandbox#

 I'm at a loss as to what to check or what do next to get past this situation. 
  Any help would be greatly appreciated.

 Thanks,

 Todd

 Todd Blackwell
 HARRIS Corporation
321-984-6911
  954-817-3662
eblac...@harris.com

 -Original Message-
 From: ceph-devel-ow...@vger.kernel.org 
 [mailto:ceph-devel-ow...@vger.kernel.org] On Behalf Of Dan Mick
 Sent: Monday, November 19, 2012 10:47 PM
 To: Blackwell, Edward
 Cc: ceph-devel@vger.kernel.org
 Subject: Re: Remote Ceph Install



 On 11/19/2012 11:42 AM, Blackwell, Edward wrote:
 Hi,
 I work for Harris Corporation, and we are investigating Ceph as a potential 
 solution to a storage problem that one of our government customers is 
 currently having.  I've already created a two-node cluster on a couple of 
 VMs with another VM acting as an administrative client.  The cluster was 
 created using some installation instructions supplied to us via Inktank, and 
 through the use of the ceph-deploy script.  Aside from a couple of quirky 
 discrepancies between the installation instructions and my environment, 
 everything went well.  My issue has cropped up on the second cluster I'm 
 trying to create, which is using a VM and a non-VM server for the nodes in 
 the cluster.  Eventually, both nodes in this cluster will be non-VMs, but 
 we're still waiting on the hardware for the second node, so I'm using a VM 
 in the meantime just to get this second cluster up and going.  Of course, 
 the administrative client node is still a VM.

 Hi Ed.  Welcome.

 The problem that I'm having with this second cluster concerns the non-VM 
 server (elsceph01 for the sake of the commands mentioned from here on out).  
 In particular, the issue crops up with the ceph-deploy install elsceph01 
 command I'm executing on my client VM (cephclient01) to install Ceph on the 
 non-VM server. The installation doesn't appear to be working as the command 
 does not return the OK message that it should when it completes 
 successfully.  I've tried using the verbose option on the command to see if 
 that sheds any light on the subject, but alas, it does not:


 root@cephclient01:~/my-admin-sandbox# ceph-deploy -v install elsceph01
 DEBUG:ceph_deploy.install:Installing stable version argonaut on cluster ceph 
 hosts elsceph01
 DEBUG:ceph_deploy.install:Detecting platform for host elsceph01 ...
 DEBUG:ceph_deploy.install:Installing for Ubuntu 12.04 on host elsceph01 ...
 root@cephclient01:~/my-admin-sandbox#


 Would you happen to have a breakdown of the commands being executed by the 
 ceph-deploy script behind the scenes so I can maybe execute them one-by-one 
 to see where the error is?  I have

Re: Remote Ceph Install

2012-12-03 Thread Dan Mick
I'm reading README.rst in the ceph-deploy sources, and I've executed the 
command, but didn't really analyze what it did when I did; I just 
assumed from that doc that it would be necessary.  It's possible your 
instructions obviate the need for that step, but I'd have to look at the 
instructions you were given, I guess...



On 12/03/2012 01:14 PM, Blackwell, Edward wrote:

Hi Dan,

In the version of the Ceph installation instructions I was given, I don't have a 
ceph-deploy gatherkeys step.  Is there a newer version, or can you briefly 
describe the use of this command?

Thanks,

Todd

Todd Blackwell
HARRIS Corporation
   321-984-6911
 954-817-3662
   eblac...@harris.com


-Original Message-
From: Dan Mick [mailto:dan.m...@inktank.com]
Sent: Monday, December 03, 2012 3:37 PM
To: Blackwell, Edward
Cc: ceph-devel@vger.kernel.org
Subject: Re: Remote Ceph Install



On 12/03/2012 10:53 AM, Blackwell, Edward wrote:

Hi Dan,

Thanks for the welcome and the advice.  There indeed was a problem with the host name and 
capitalization as you described, but once I corrected that, a new issue began to occur 
when I ran the ceph-deploy mon command.  The command appears to run 
successfully (no output from the command is generated), but when I check the status on 
one of the servers in the cluster (ceph04 and ELSCEPH01) as recommended by the 
directions, I get the following:

root@cephclient01:~/my-admin-sandbox# ceph-deploy mon
root@cephclient01:~/my-admin-sandbox# ssh ceph04 ceph -s
2012-12-03 13:26:30.031854 7fc353d60780 -1 auth: failed to open keyring from 
/etc/ceph/ceph.client.admin.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin
2012-12-03 13:26:30.031963 7fc353d60780 -1 monclient(hunting): failed to open 
keyring: (2) No such file or directory
2012-12-03 13:26:30.032042 7fc353d60780 -1 ceph_tool_common_init failed.
root@cephclient01:~/my-admin-sandbox#


Yeah, looks like the keys didn't get distributed correctly.  Did you do
the ceph-deploy gatherkeys step?



Behind the scenes, is the ceph-deploy mon command executing the mkcephfs 
command, which creates the keyring file?  If so, could that command be failing somehow, 
and hence the status command is not able to return the status of the Ceph installation?

I even tried executing the ceph-deploy mon command using the -v option, and 
got the following, so it seems to be working correctly:

root@cephclient01:~/my-admin-sandbox# ceph-deploy -v mon
DEBUG:ceph_deploy.mon:Deploying mon, cluster ceph hosts ceph04 ELSCEPH01
DEBUG:ceph_deploy.mon:Deploying mon to ceph04
DEBUG:ceph_deploy.mon:Deploying mon to ELSCEPH01
root@cephclient01:~/my-admin-sandbox#

I'm at a loss as to what to check or what do next to get past this situation.  
Any help would be greatly appreciated.

Thanks,

Todd

Todd Blackwell
HARRIS Corporation
321-984-6911
 954-817-3662
eblac...@harris.com

-Original Message-
From: ceph-devel-ow...@vger.kernel.org 
[mailto:ceph-devel-ow...@vger.kernel.org] On Behalf Of Dan Mick
Sent: Monday, November 19, 2012 10:47 PM
To: Blackwell, Edward
Cc: ceph-devel@vger.kernel.org
Subject: Re: Remote Ceph Install



On 11/19/2012 11:42 AM, Blackwell, Edward wrote:

Hi,
I work for Harris Corporation, and we are investigating Ceph as a potential 
solution to a storage problem that one of our government customers is currently 
having.  I've already created a two-node cluster on a couple of VMs with 
another VM acting as an administrative client.  The cluster was created using 
some installation instructions supplied to us via Inktank, and through the use 
of the ceph-deploy script.  Aside from a couple of quirky discrepancies between 
the installation instructions and my environment, everything went well.  My 
issue has cropped up on the second cluster I'm trying to create, which is using 
a VM and a non-VM server for the nodes in the cluster.  Eventually, both nodes 
in this cluster will be non-VMs, but we're still waiting on the hardware for 
the second node, so I'm using a VM in the meantime just to get this second 
cluster up and going.  Of course, the administrative client node is still a VM.


Hi Ed.  Welcome.


The problem that I'm having with this second cluster concerns the non-VM server 
(elsceph01 for the sake of the commands mentioned from here on out).  In 
particular, the issue crops up with the ceph-deploy install elsceph01 command 
I'm executing on my client VM (cephclient01) to install Ceph on the non-VM 
server. The installation doesn't appear to be working as the command does not 
return the OK message that it should when it completes successfully.  I've 
tried using the verbose option on the command to see if that sheds any light on 
the subject, but alas, it does not:


root@cephclient01:~/my-admin-sandbox# ceph-deploy -v install elsceph01
DEBUG:ceph_deploy.install:Installing stable version argonaut on cluster ceph 
hosts elsceph01
DEBUG:ceph_deploy.install:Detecting platform for host elsceph01

Remote Ceph Install

2012-11-19 Thread Blackwell, Edward
Hi,
I work for Harris Corporation, and we are investigating Ceph as a potential 
solution to a storage problem that one of our government customers is currently 
having.  I've already created a two-node cluster on a couple of VMs with 
another VM acting as an administrative client.  The cluster was created using 
some installation instructions supplied to us via Inktank, and through the use 
of the ceph-deploy script.  Aside from a couple of quirky discrepancies between 
the installation instructions and my environment, everything went well.  My 
issue has cropped up on the second cluster I'm trying to create, which is using 
a VM and a non-VM server for the nodes in the cluster.  Eventually, both nodes 
in this cluster will be non-VMs, but we're still waiting on the hardware for 
the second node, so I'm using a VM in the meantime just to get this second 
cluster up and going.  Of course, the administrative client node is still a VM.

The problem that I'm having with this second cluster concerns the non-VM server 
(elsceph01 for the sake of the commands mentioned from here on out).  In 
particular, the issue crops up with the ceph-deploy install elsceph01 command 
I'm executing on my client VM (cephclient01) to install Ceph on the non-VM 
server. The installation doesn't appear to be working as the command does not 
return the OK message that it should when it completes successfully.  I've 
tried using the verbose option on the command to see if that sheds any light on 
the subject, but alas, it does not:


root@cephclient01:~/my-admin-sandbox# ceph-deploy -v install elsceph01
DEBUG:ceph_deploy.install:Installing stable version argonaut on cluster ceph 
hosts elsceph01
DEBUG:ceph_deploy.install:Detecting platform for host elsceph01 ...
DEBUG:ceph_deploy.install:Installing for Ubuntu 12.04 on host elsceph01 ...
root@cephclient01:~/my-admin-sandbox#


Would you happen to have a breakdown of the commands being executed by the 
ceph-deploy script behind the scenes so I can maybe execute them one-by-one to 
see where the error is?  I have confirmed that it looks like the installation 
of the software has succeeded as I did a which ceph command on elsceph01, and 
it reported back /usr/bin/ceph.  Also, /etc/ceph/ceph.conf is there, and it 
matches the file created by the ceph-deploy new ... command on the client.  
Does the install command do a mkcephfs behind the scenes?  The reason I ask is 
that when I do the ceph-deploy mon command from the client, which is the next 
command listed in the instructions to do, I get this output:


root@cephclient01:~/my-admin-sandbox# ceph-deploy mon
creating /var/lib/ceph/tmp/ceph-ELSCEPH01.mon.keyring
2012-11-15 11:35:38.954261 7f7a6c274780 -1 asok(0x260b000) 
AdminSocketConfigObs::init: failed: AdminSocket::bind_and_listen: failed to 
bind the UNIX domain socket to '/var/run/ceph/ceph-mon.ELSCEPH01.asok': (2) No 
such file or directory
Traceback (most recent call last):
  File /usr/local/bin/ceph-deploy, line 9, in module
load_entry_point('ceph-deploy==0.0.1', 'console_scripts', 'ceph-deploy')()
  File /root/ceph-deploy/ceph_deploy/cli.py, line 80, in main
added entity mon. auth auth(auid = 18446744073709551615 
key=AQBWDj5QAP6LHhAAskVBnUkYHJ7eYREmKo5qKA== with 0 caps)
return args.func(args)
mon/MonMap.h: In function 'void MonMap::add(const string, const 
entity_addr_t)' thread 7f7a6c274780 time 2012-11-15 11:35:38.955024
mon/MonMap.h: 97: FAILED assert(addr_name.count(addr) == 0)
ceph version 0.48.2argonaut (commit:3e02b2fad88c2a95d9c0c86878f10d1beb780bfe)
1: (MonMap::build_from_host_list(std::string, std::string)+0x738) [0x5988b8]
2: (MonMap::build_initial(CephContext*, std::ostream)+0x113) [0x59bd53]
3: (main()+0x12bb) [0x45ffab]
4: (__libc_start_main()+0xed) [0x7f7a6a6d776d]
5: ceph-mon() [0x462a19]
NOTE: a copy of the executable, or `objdump -rdS executable` is needed to 
interpret this.
2012-11-15 11:35:38.955924 7f7a6c274780 -1 mon/MonMap.h: In function 'void 
MonMap::add(const string, const entity_addr_t)' thread 7f7a6c274780 time 
2012-11-15 11:35:38.955024
mon/MonMap.h: 97: FAILED assert(addr_name.count(addr) == 0)

ceph version 0.48.2argonaut (commit:3e02b2fad88c2a95d9c0c86878f10d1beb780bfe)
1: (MonMap::build_from_host_list(std::string, std::string)+0x738) [0x5988b8]
2: (MonMap::build_initial(CephContext*, std::ostream)+0x113) [0x59bd53]
3: (main()+0x12bb) [0x45ffab]
4: (__libc_start_main()+0xed) [0x7f7a6a6d776d]
5: ceph-mon() [0x462a19]
NOTE: a copy of the executable, or `objdump -rdS executable` is needed to 
interpret this.

-1 2012-11-15 11:35:38.954261 7f7a6c274780 -1 asok(0x260b000) 
AdminSocketConfigObs::init: failed: AdminSocket::bind_and_listen: failed to 
bind the UNIX domain socket to '/var/run/ceph/ceph-mon.ELSCEPH01.asok': (2) No 
such file or directory
 0 2012-11-15 11:35:38.955924 7f7a6c274780 -1 mon/MonMap.h: In function 
'void MonMap::add(const string, const entity_addr_t)' thread 7f7a6c274780 
time 2012-11-15 11:35:38.955024

Re: Remote Ceph Install

2012-11-19 Thread Dan Mick



On 11/19/2012 11:42 AM, Blackwell, Edward wrote:

Hi,
I work for Harris Corporation, and we are investigating Ceph as a potential 
solution to a storage problem that one of our government customers is currently 
having.  I've already created a two-node cluster on a couple of VMs with 
another VM acting as an administrative client.  The cluster was created using 
some installation instructions supplied to us via Inktank, and through the use 
of the ceph-deploy script.  Aside from a couple of quirky discrepancies between 
the installation instructions and my environment, everything went well.  My 
issue has cropped up on the second cluster I'm trying to create, which is using 
a VM and a non-VM server for the nodes in the cluster.  Eventually, both nodes 
in this cluster will be non-VMs, but we're still waiting on the hardware for 
the second node, so I'm using a VM in the meantime just to get this second 
cluster up and going.  Of course, the administrative client node is still a VM.


Hi Ed.  Welcome.


The problem that I'm having with this second cluster concerns the non-VM server 
(elsceph01 for the sake of the commands mentioned from here on out).  In 
particular, the issue crops up with the ceph-deploy install elsceph01 command 
I'm executing on my client VM (cephclient01) to install Ceph on the non-VM 
server. The installation doesn't appear to be working as the command does not 
return the OK message that it should when it completes successfully.  I've 
tried using the verbose option on the command to see if that sheds any light on 
the subject, but alas, it does not:


root@cephclient01:~/my-admin-sandbox# ceph-deploy -v install elsceph01
DEBUG:ceph_deploy.install:Installing stable version argonaut on cluster ceph 
hosts elsceph01
DEBUG:ceph_deploy.install:Detecting platform for host elsceph01 ...
DEBUG:ceph_deploy.install:Installing for Ubuntu 12.04 on host elsceph01 ...
root@cephclient01:~/my-admin-sandbox#


Would you happen to have a breakdown of the commands being executed by the 
ceph-deploy script behind the scenes so I can maybe execute them one-by-one to 
see where the error is?  I have confirmed that it looks like the installation 
of the software has succeeded as I did a which ceph command on elsceph01, and 
it reported back /usr/bin/ceph.  Also, /etc/ceph/ceph.conf is there, and it 
matches the file created by the ceph-deploy new ... command on the client.  
Does the install command do a mkcephfs behind the scenes?  The reason I ask is 
that when I do the ceph-deploy mon command from the client, which is the next 
command listed in the instructions to do, I get this output:


Basically install just runs the appropriate debian package commands to 
get the requested release of Ceph installed on the target host (in this 
case, defaulting to argonaut).  The command normally doesn't issue any 
output.



root@cephclient01:~/my-admin-sandbox# ceph-deploy mon
creating /var/lib/ceph/tmp/ceph-ELSCEPH01.mon.keyring


This looks like there may be confusion about case in the hostname.  What 
does hostname on elsceph01 report?  If it's ELSCEPH01, that's probably 
the problem; the pathnames etc. are all case-sensitive.
Could be that /etc/hosts has the wrong case, or both cases, of the 
hostname in it?



2012-11-15 11:35:38.954261 7f7a6c274780 -1 asok(0x260b000) 
AdminSocketConfigObs::init: failed: AdminSocket::bind_and_listen: failed to 
bind the UNIX domain socket to '/var/run/ceph/ceph-mon.ELSCEPH01.asok': (2) No 
such file or directory
Traceback (most recent call last):
   File /usr/local/bin/ceph-deploy, line 9, in module
 load_entry_point('ceph-deploy==0.0.1', 'console_scripts', 'ceph-deploy')()
   File /root/ceph-deploy/ceph_deploy/cli.py, line 80, in main
added entity mon. auth auth(auid = 18446744073709551615 
key=AQBWDj5QAP6LHhAAskVBnUkYHJ7eYREmKo5qKA== with 0 caps)
 return args.func(args)
mon/MonMap.h: In function 'void MonMap::add(const string, const 
entity_addr_t)' thread 7f7a6c274780 time 2012-11-15 11:35:38.955024
mon/MonMap.h: 97: FAILED assert(addr_name.count(addr) == 0)
ceph version 0.48.2argonaut (commit:3e02b2fad88c2a95d9c0c86878f10d1beb780bfe)
1: (MonMap::build_from_host_list(std::string, std::string)+0x738) [0x5988b8]
2: (MonMap::build_initial(CephContext*, std::ostream)+0x113) [0x59bd53]
3: (main()+0x12bb) [0x45ffab]
4: (__libc_start_main()+0xed) [0x7f7a6a6d776d]
5: ceph-mon() [0x462a19]
NOTE: a copy of the executable, or `objdump -rdS executable` is needed to 
interpret this.
2012-11-15 11:35:38.955924 7f7a6c274780 -1 mon/MonMap.h: In function 'void 
MonMap::add(const string, const entity_addr_t)' thread 7f7a6c274780 time 
2012-11-15 11:35:38.955024
mon/MonMap.h: 97: FAILED assert(addr_name.count(addr) == 0)

ceph version 0.48.2argonaut (commit:3e02b2fad88c2a95d9c0c86878f10d1beb780bfe)
1: (MonMap::build_from_host_list(std::string, std::string)+0x738) [0x5988b8]
2: (MonMap::build_initial(CephContext*, std::ostream)+0x113) [0x59bd53]
3: (main()+0x12bb) [0x45ffab]
4: