Re: [ceph-users] firefly osds stuck in state booting

2014-07-29 Thread 10 minus
Hi Karan ,

Thanks .. that did the trick ..
The magic word was "in"
regarding rep size . I have adjusted them
my settings are
--snip--
osd pool default size =
2

osd pool default min size =
1

osd pool default pg num =
100

osd pool default pgp num = 100
--snip--

# Also in the meantime I had chance to play with ceph-deploy script too.
# Maybe it was me or probably it is a bug . I have tried twice and
everytime I have hit this

As I said before I'm using a directory as this is a test installation .

ceph-deploy osd prepare ceph2:/ceph2:/ceph2/journald <=== Works


but

--snip--
ceph-deploy osd activate
ceph2:/ceph2:/ceph2/journald

[ceph_deploy.conf][DEBUG ] found configuration file at:
/home/ceph/.cephdeploy.conf
[ceph_deploy.cli][INFO  ] Invoked (1.5.9): /usr/bin/ceph-deploy osd
activate ceph2:/ceph2:/ceph2/journald
[ceph_deploy.osd][DEBUG ] Activating cluster ceph disks
ceph2:/ceph2:/ceph2/journald
[ceph2][DEBUG ] connected to host: ceph2
[ceph2][DEBUG ] detect platform information from remote host
[ceph2][DEBUG ] detect machine type
[ceph_deploy.osd][INFO  ] Distro info: CentOS 6.5 Final
[ceph_deploy.osd][DEBUG ] activating host ceph2 disk /ceph2
[ceph_deploy.osd][DEBUG ] will use init type: sysvinit
[ceph2][INFO  ] Running command: sudo ceph-disk-activate --mark-init
sysvinit --mount /ceph2
[ceph2][WARNIN] got monmap epoch 2
[ceph2][WARNIN] 2014-07-28 11:47:04.733204 7f08d1c667a0 -1 journal
FileJournal::_open: disabling aio for non-block journal.  Use
journal_force_aio to force
 use of aio anyway
[ceph2][WARNIN] 2014-07-28 11:47:04.733400 7f08d1c667a0 -1 journal check:
ondisk fsid ---- doesn't match expected
4795daff-
d63f-415b-9824-75f0863eb14f, invalid (someone else's?) journal
[ceph2][WARNIN] 2014-07-28 11:47:04.796835 7f08d1c667a0 -1 journal
FileJournal::_open: disabling aio for non-block journal.  Use
journal_force_aio to force
 use of aio anyway
[ceph2][WARNIN] 2014-07-28 11:47:04.798944 7f08d1c667a0 -1
filestore(/ceph2) could not find 23c2fcde/osd_superblock/0//-1 in index:
(2) No such file or dir
ectory
[ceph2][WARNIN] 2014-07-28 11:47:04.874282 7f08d1c667a0 -1 created object
store /ceph2 journal /ceph2/journal for osd.1 fsid
109507ab-adf1-4eb6-aacf-092549
4e3882
[ceph2][WARNIN] 2014-07-28 11:47:04.874474 7f08d1c667a0 -1 auth: error
reading file: /ceph2/keyring: can't open /ceph2/keyring: (2) No such file
or directo
ry
[ceph2][WARNIN] 2014-07-28 11:47:04.875209 7f08d1c667a0 -1 created new key
in keyring /ceph2/keyring
[ceph2][WARNIN] added key for osd.1
[ceph2][WARNIN] ceph-disk: Error: unable to create symlink
/var/lib/ceph/osd/ceph-1 -> /ceph2
[ceph2][ERROR ] RuntimeError: command returned non-zero exit status: 1
[ceph_deploy][ERROR ] RuntimeError: Failed to execute command:
ceph-disk-activate --mark-init sysvinit --mount /ceph2
--snip--



It turns out ceph-deploy does not create the directory /var/lib/ceph/osd
and if I create them everything works.


Cheers


On Mon, Jul 28, 2014 at 9:09 AM, Karan Singh  wrote:

> The output that you have provided says that OSDs are not IN , Try the below
>
> ceph osd in osd.0
> ceph osd in osd.1
>
> service ceph start osd.0
> service ceph start osd.1
>
> If you have 1 more host with 1 disk , add it , starting Ceph Firefly
> default rep size is 3
>
>
> - Karan -
>
> On 27 Jul 2014, at 11:17, 10 minus  wrote:
>
> Hi Sage,
>
> I have dropped all unset .. and even restarted the osd
> No dice .. OSDs are still stuck .
>
>
>
> --snip--
>  ceph daemon osd.0 status│rtt
> min/avg/max/mdev = 0.095/0.120/0.236/0.015
> ms
>
> { "cluster_fsid":
> "99babb8f-c880-4b32-a227-94aa483d4871",
> │[root@ceph2 ~]#  ceph daemon osd.1
> status
>
>   "osd_fsid":
> "1ad28bde-c23c-44ba-a3b7-0fd3372e",│{
> "cluster_fsid":
> "99babb8f-c880-4b32-a227-94aa483d4871",
>
>   "whoami":
> 0,   │
> "osd_fsid":
> "becc3252-6977-47d6-87af-7b1337e591d8",
>
>   "state":
> "booting",
> │  "whoami":
> 1,
>
>   "oldest_map":
> 1,   │
> "state":
> "booting",
>
>   "newest_map":
> 24,  │
> "oldest_map":
> 1,
>
>   "num_pgs":
> 0}  │
> "newest_map":
> 21,
>
>  --snip--
>
> --snip--
> ceph osd
> tree
>
> # idweight  type name   up/down reweight
> -1  2   root default
> -3  1   host ceph1
> 0   1   osd.0   down0
> -2  1   host ceph2
> 1   1   osd.1   down0
>
>  --snip--
>
> --snip--
>  ceph -s
> cluster 2929fa80-0841-4cb6-a133-90b2098fc802
>  health HEALTH_WARN 192 pgs stuck inactive; 192 pgs stuck unclean
>  monmap e2: 3 mons at {ceph0=
> 10.0.12.220:6789/0,ceph1=10.0.12.221:6789/0,ceph2=10.0.12.222:6789/0},
> election e

Re: [ceph-users] firefly osds stuck in state booting

2014-07-28 Thread Karan Singh
The output that you have provided says that OSDs are not IN , Try the below

ceph osd in osd.0
ceph osd in osd.1

service ceph start osd.0
service ceph start osd.1

If you have 1 more host with 1 disk , add it , starting Ceph Firefly default 
rep size is 3


- Karan -

On 27 Jul 2014, at 11:17, 10 minus  wrote:

> Hi Sage, 
> 
> I have dropped all unset .. and even restarted the osd 
> No dice .. OSDs are still stuck .
> 
> 
> 
> --snip--
>  ceph daemon osd.0 status│rtt 
> min/avg/max/mdev = 0.095/0.120/0.236/0.015 ms 
> 
> { "cluster_fsid": "99babb8f-c880-4b32-a227-94aa483d4871", 
>│[root@ceph2 ~]#  ceph daemon osd.1 status 
>   
>   
>   "osd_fsid": "1ad28bde-c23c-44ba-a3b7-0fd3372e", 
>│{ "cluster_fsid": "99babb8f-c880-4b32-a227-94aa483d4871", 
>   
>   
>   "whoami": 0,
>│  "osd_fsid": "becc3252-6977-47d6-87af-7b1337e591d8", 
>   
>   
>   "state": "booting", 
>│  "whoami": 1,
>   
>   
>   "oldest_map": 1,
>│  "state": "booting", 
>   
>   
>   "newest_map": 24,   
>│  "oldest_map": 1,
>   
>   
>   "num_pgs": 0}   
>│  "newest_map": 21,   
>   
>   
>  --snip--
> 
> --snip-- 
> ceph osd tree 
>  
> # idweight  type name   up/down reweight
> -1  2   root default
> -3  1   host ceph1
> 0   1   osd.0   down0
> -2  1   host ceph2
> 1   1   osd.1   down0
> 
>  --snip--
> 
> --snip--
>  ceph -s
> cluster 2929fa80-0841-4cb6-a133-90b2098fc802
>  health HEALTH_WARN 192 pgs stuck inactive; 192 pgs stuck unclean
>  monmap e2: 3 mons at 
> {ceph0=10.0.12.220:6789/0,ceph1=10.0.12.221:6789/0,ceph2=10.0.12.222:6789/0}, 
> election epoch 50, quorum 0,1,2 ceph0,ceph1,ceph2
>  osdmap e24: 2 osds: 0 up, 0 in
>   pgmap v25: 192 pgs, 3 pools, 0 bytes data, 0 objects
> 0 kB used, 0 kB / 0 kB avail
>  192 creating
> --snip--
> 
> 
> 
> 
> On Sat, Jul 26, 2014 at 5:57 PM, Sage Weil  wrote:
> On Sat, 26 Jul 2014, 10 minus wrote:
> > Hi,
> >
> > I just setup a test ceph installation on 3 node Centos 6.5  .
> > two of the nodes are used for hosting osds and the third acts as mon .
> >
> > Please note I'm using LVM so had to set up the osd using the manual install
> > guide.
> >
> > --snip--
> > ceph -s
> > cluster 2929fa80-0841-4cb6-a133-90b2098fc802
> >  health HEALTH_WARN 192 pgs stuck inactive; 192 pgs stuck unclean;
> > noup,nodown,noout flag(s) set
> >  monmap e2: 3 mons 
> > at{ceph0=10.0.12.220:6789/0,ceph1=10.0.12.221:6789/0,ceph2=10.0.12.222:6789/0
> > }, election epoch 46, quorum 0,1,2 ceph0,ceph1,ceph2
> >  osdmap e21: 2 osds: 0 up, 0 in
> > flags noup,nodown,noout
> 
> 
> Do 'ceph osd unset noup' and they should start up.  You likely also want
> to clear nodown and noout as well.
> 
> sage
> 
> 
> >   pgmap v22: 192 pgs, 3 pools, 0 bytes data, 0 objects
> > 0 kB used, 0 kB / 0 kB avail
> >  192 creating
> > --snip--
> >
> > osd tree
> >
> > --snip--
> > ceph osd tree
> > # idweight  type name   up/down reweight
> > -1  2   root default
> > -3  1   host ceph1
> > 0   1   osd.0   down0
> > -2  1   host ceph2
> > 1   1   osd.1   down0
> > --snip--
> >
> > --snip--
> >  ceph daemon osd.0 status
> > { "cluster_fsid": "99babb8f-c880-4b32-a227-94aa483d4871",
> >   "osd_fsid": "1ad28bde-c23c-44ba-a3b7-0fd3372e",
> >   "whoami": 0,
> >   "state": "booting",
> >  

Re: [ceph-users] firefly osds stuck in state booting

2014-07-27 Thread 10 minus
Hi Sage,

I have dropped all unset .. and even restarted the osd
No dice .. OSDs are still stuck .



--snip--
 ceph daemon osd.0 status│rtt
min/avg/max/mdev = 0.095/0.120/0.236/0.015
ms

{ "cluster_fsid":
"99babb8f-c880-4b32-a227-94aa483d4871",
│[root@ceph2 ~]#  ceph daemon osd.1
status

  "osd_fsid":
"1ad28bde-c23c-44ba-a3b7-0fd3372e",│{
"cluster_fsid":
"99babb8f-c880-4b32-a227-94aa483d4871",

  "whoami":
0,   │
"osd_fsid":
"becc3252-6977-47d6-87af-7b1337e591d8",

  "state":
"booting",
│  "whoami":
1,

  "oldest_map":
1,   │
"state":
"booting",

  "newest_map":
24,  │
"oldest_map":
1,

  "num_pgs":
0}  │
"newest_map":
21,

 --snip--

--snip--
ceph osd
tree

# idweight  type name   up/down reweight
-1  2   root default
-3  1   host ceph1
0   1   osd.0   down0
-2  1   host ceph2
1   1   osd.1   down0

 --snip--

--snip--
 ceph -s
cluster 2929fa80-0841-4cb6-a133-90b2098fc802
 health HEALTH_WARN 192 pgs stuck inactive; 192 pgs stuck unclean
 monmap e2: 3 mons at {ceph0=
10.0.12.220:6789/0,ceph1=10.0.12.221:6789/0,ceph2=10.0.12.222:6789/0},
election epoch 50, quorum 0,1,2 ceph0,ceph1,ceph2
 osdmap e24: 2 osds: 0 up, 0 in
  pgmap v25: 192 pgs, 3 pools, 0 bytes data, 0 objects
0 kB used, 0 kB / 0 kB avail
 192 creating
--snip--




On Sat, Jul 26, 2014 at 5:57 PM, Sage Weil  wrote:

> On Sat, 26 Jul 2014, 10 minus wrote:
> > Hi,
> >
> > I just setup a test ceph installation on 3 node Centos 6.5  .
> > two of the nodes are used for hosting osds and the third acts as mon .
> >
> > Please note I'm using LVM so had to set up the osd using the manual
> install
> > guide.
> >
> > --snip--
> > ceph -s
> > cluster 2929fa80-0841-4cb6-a133-90b2098fc802
> >  health HEALTH_WARN 192 pgs stuck inactive; 192 pgs stuck unclean;
> > noup,nodown,noout flag(s) set
> >  monmap e2: 3 mons at{ceph0=
> 10.0.12.220:6789/0,ceph1=10.0.12.221:6789/0,ceph2=10.0.12.222:6789/0
> > }, election epoch 46, quorum 0,1,2 ceph0,ceph1,ceph2
> >  osdmap e21: 2 osds: 0 up, 0 in
> > flags noup,nodown,noout
> 
>
> Do 'ceph osd unset noup' and they should start up.  You likely also want
> to clear nodown and noout as well.
>
> sage
>
>
> >   pgmap v22: 192 pgs, 3 pools, 0 bytes data, 0 objects
> > 0 kB used, 0 kB / 0 kB avail
> >  192 creating
> > --snip--
> >
> > osd tree
> >
> > --snip--
> > ceph osd tree
> > # idweight  type name   up/down reweight
> > -1  2   root default
> > -3  1   host ceph1
> > 0   1   osd.0   down0
> > -2  1   host ceph2
> > 1   1   osd.1   down0
> > --snip--
> >
> > --snip--
> >  ceph daemon osd.0 status
> > { "cluster_fsid": "99babb8f-c880-4b32-a227-94aa483d4871",
> >   "osd_fsid": "1ad28bde-c23c-44ba-a3b7-0fd3372e",
> >   "whoami": 0,
> >   "state": "booting",
> >   "oldest_map": 1,
> >   "newest_map": 21,
> >   "num_pgs": 0}
> >
> > --snip--
> >
> > --snip--
> >  ceph daemon osd.1 status
> > { "cluster_fsid": "99babb8f-c880-4b32-a227-94aa483d4871",
> >   "osd_fsid": "becc3252-6977-47d6-87af-7b1337e591d8",
> >   "whoami": 1,
> >   "state": "booting",
> >   "oldest_map": 1,
> >   "newest_map": 21,
> >   "num_pgs": 0}
> > --snip--
> >
> > # Cpus are idling
> >
> > # does anybody know what is wrong
> >
> > Thanks in advance
> >
> >
>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] firefly osds stuck in state booting

2014-07-26 Thread Sage Weil
On Sat, 26 Jul 2014, 10 minus wrote:
> Hi,
> 
> I just setup a test ceph installation on 3 node Centos 6.5  .
> two of the nodes are used for hosting osds and the third acts as mon .
> 
> Please note I'm using LVM so had to set up the osd using the manual install
> guide.
> 
> --snip--
> ceph -s
>     cluster 2929fa80-0841-4cb6-a133-90b2098fc802
>  health HEALTH_WARN 192 pgs stuck inactive; 192 pgs stuck unclean;
> noup,nodown,noout flag(s) set
>  monmap e2: 3 mons 
> at{ceph0=10.0.12.220:6789/0,ceph1=10.0.12.221:6789/0,ceph2=10.0.12.222:6789/0
> }, election epoch 46, quorum 0,1,2 ceph0,ceph1,ceph2
>  osdmap e21: 2 osds: 0 up, 0 in
>     flags noup,nodown,noout


Do 'ceph osd unset noup' and they should start up.  You likely also want 
to clear nodown and noout as well.

sage


>   pgmap v22: 192 pgs, 3 pools, 0 bytes data, 0 objects
>     0 kB used, 0 kB / 0 kB avail
>  192 creating
> --snip--
> 
> osd tree
> 
> --snip--
> ceph osd tree
> # id    weight  type name   up/down reweight
> -1  2   root default
> -3  1   host ceph1
> 0   1   osd.0   down    0
> -2  1   host ceph2
> 1   1   osd.1   down    0
> --snip--
> 
> --snip--
>  ceph daemon osd.0 status
> { "cluster_fsid": "99babb8f-c880-4b32-a227-94aa483d4871",
>   "osd_fsid": "1ad28bde-c23c-44ba-a3b7-0fd3372e",
>   "whoami": 0,
>   "state": "booting",
>   "oldest_map": 1,
>   "newest_map": 21,
>   "num_pgs": 0}
> 
> --snip--
> 
> --snip--
>  ceph daemon osd.1 status
> { "cluster_fsid": "99babb8f-c880-4b32-a227-94aa483d4871",
>   "osd_fsid": "becc3252-6977-47d6-87af-7b1337e591d8",
>   "whoami": 1,
>   "state": "booting",
>   "oldest_map": 1,
>   "newest_map": 21,
>   "num_pgs": 0}
> --snip--
> 
> # Cpus are idling
> 
> # does anybody know what is wrong
> 
> Thanks in advance
> 
> ___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] firefly osds stuck in state booting

2014-07-26 Thread 10 minus
Hi,

I just setup a test ceph installation on 3 node Centos 6.5  .
two of the nodes are used for hosting osds and the third acts as mon .

Please note I'm using LVM so had to set up the osd using the manual install
guide.

--snip--
ceph -s
cluster 2929fa80-0841-4cb6-a133-90b2098fc802
 health HEALTH_WARN 192 pgs stuck inactive; 192 pgs stuck unclean;
noup,nodown,noout flag(s) set
 monmap e2: 3 mons at {ceph0=
10.0.12.220:6789/0,ceph1=10.0.12.221:6789/0,ceph2=10.0.12.222:6789/0},
election epoch 46, quorum 0,1,2 ceph0,ceph1,ceph2
 osdmap e21: 2 osds: 0 up, 0 in
flags noup,nodown,noout
  pgmap v22: 192 pgs, 3 pools, 0 bytes data, 0 objects
0 kB used, 0 kB / 0 kB avail
 192 creating
--snip--

osd tree

--snip--
ceph osd tree
# idweight  type name   up/down reweight
-1  2   root default
-3  1   host ceph1
0   1   osd.0   down0
-2  1   host ceph2
1   1   osd.1   down0
--snip--

--snip--
 ceph daemon osd.0 status
{ "cluster_fsid": "99babb8f-c880-4b32-a227-94aa483d4871",
  "osd_fsid": "1ad28bde-c23c-44ba-a3b7-0fd3372e",
  "whoami": 0,
  "state": "booting",
  "oldest_map": 1,
  "newest_map": 21,
  "num_pgs": 0}

--snip--

--snip--
 ceph daemon osd.1 status
{ "cluster_fsid": "99babb8f-c880-4b32-a227-94aa483d4871",
  "osd_fsid": "becc3252-6977-47d6-87af-7b1337e591d8",
  "whoami": 1,
  "state": "booting",
  "oldest_map": 1,
  "newest_map": 21,
  "num_pgs": 0}
--snip--

# Cpus are idling

# does anybody know what is wrong

Thanks in advance
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com