[ceph-users] "unable to connect to cluster" after monitor IP change

2018-04-06 Thread Nathan Dehnel
gentooserver ~ # ceph-mon -i mon0 --extract-monmap /tmp/monmap
2018-04-06 15:38:10.863444 7f8aa2b72f80 -1 wrote monmap to /tmp/monmap
gentooserver ~ # monmaptool --print /tmp/monmap
monmaptool: monmap file /tmp/monmap
epoch 3
fsid a736559a-92d1-483e-9289-d2c7feed510f
last_changed 2018-04-06 14:53:12.892574
created 2018-04-06 14:52:18.190509
0: [2605:6000:1020:2056:7d79:3f08:ee64:2aa3]:6789/0 mon.mon0

This is the monmap I injected into my monitor.


gentooserver ~ # systemctl status ceph-mon@mon0
● ceph-mon@mon0.service - Ceph cluster monitor daemon
   Loaded: loaded (/lib/systemd/system/ceph-mon@.service; indirect; vendor
preset: disabled)
   Active: active (running) since Fri 2018-04-06 15:47:51 CDT; 16s ago
 Main PID: 4362 (ceph-mon)
   CGroup: /system.slice/system-ceph\x2dmon.slice/ceph-mon@mon0.service
   └─4362 /usr/bin/ceph-mon -f --cluster ceph --id mon0 --setuser
ceph --setgroup ceph

Apr 06 15:47:51 gentooserver systemd[1]: Started Ceph cluster monitor
daemon.
Apr 06 15:47:51 gentooserver ceph-mon[4362]: 2018-04-06 15:47:51.841218
7f5824d44f80 -1 distro_detect - can't detect distro_version

gentooserver ~ # systemctl status ceph-mgr@mgr0
● ceph-mgr@mgr0.service - Ceph cluster manager daemon
   Loaded: loaded (/lib/systemd/system/ceph-mgr@.service; indirect; vendor
preset: disabled)
   Active: active (running) since Fri 2018-04-06 15:10:01 CDT; 38min ago
 Main PID: 3807 (ceph-mgr)
   CGroup: /system.slice/system-ceph\x2dmgr.slice/ceph-mgr@mgr0.service
   └─3807 /usr/bin/ceph-mgr -f --cluster ceph --id mgr0 --setuser
ceph --setgroup ceph

Apr 06 15:10:01 gentooserver systemd[1]: Started Ceph cluster manager
daemon.

gentooserver ~ # systemctl status ceph-mds@mds0
● ceph-mds@mds0.service - Ceph metadata server daemon
   Loaded: loaded (/lib/systemd/system/ceph-mds@.service; disabled; vendor
preset: disabled)
  Drop-In: /etc/systemd/system/ceph-mds@.service.d
   └─00gentoo.conf
   Active: active (running) since Fri 2018-04-06 15:10:25 CDT; 38min ago
 Main PID: 3827 (ceph-mds)
   CGroup: /system.slice/system-ceph\x2dmds.slice/ceph-mds@mds0.service
   └─3827 /usr/bin/ceph-mds -f --cluster ceph --id mds0 --setuser
ceph --setgroup ceph

Apr 06 15:10:25 gentooserver systemd[1]: Started Ceph metadata server
daemon.
Apr 06 15:10:25 gentooserver ceph-mds[3827]: starting mds.mds0 at -

All the daemons report they are running.


gentooserver ~ # ceph daemon mon.mon0 mon_status
{
"name": "mon0",
"rank": 0,
"state": "leader",
"election_epoch": 39,
"quorum": [
0
],
"features": {
"required_con": "153140804152475648",
"required_mon": [
"kraken",
"luminous"
],
"quorum_con": "2305244844532236283",
"quorum_mon": [
"kraken",
"luminous"
]
},
"outside_quorum": [],
"extra_probe_peers": [],
"sync_provider": [],
"monmap": {
"epoch": 3,
"fsid": "a736559a-92d1-483e-9289-d2c7feed510f",
"modified": "2018-04-06 14:53:12.892574",
"created": "2018-04-06 14:52:18.190509",
"features": {
"persistent": [
"kraken",
"luminous"
],
"optional": []
},
"mons": [
{
"rank": 0,
"name": "mon0",
"addr": "[2605:6000:1020:2056:7d79:3f08:ee64:2aa3]:6789/0",
"public_addr":
"[2605:6000:1020:2056:7d79:3f08:ee64:2aa3]:6789/0"
}
]
},
"feature_map": {
"mon": {
"group": {
"features": "0x1ffddff8eea4fffb",
"release": "luminous",
"num": 1
}
}
}
}


gentooserver ~ # cat /etc/ceph/ceph.conf
[global]
cluster = ceph
fsid = a736559a-92d1-483e-9289-d2c7feed510f
ms bind ipv6 = true
auth cluster required = none
auth service required = none
auth client required = none
mon host = gentooserver
mon addr = [2605:6000:1020:2056:7d79:3f08:ee64:2aa3]:6789
mon pg warn max per osd = 300

[mon]
mon initial members = mon0
mon host = gentooserver
mon addr = [2605:6000:1020:2056:7d79:3f08:ee64:2aa3]:6789
mon pg warn max per osd = 300
mon allow pool delete = true

[mon.mon0]
host = gentooserver
mon addr = [2605:6000:1020:2056:7d79:3f08:ee64:2aa3]:6789

[osd]
osd journal size = 1
osd crush chooseleaf type = 0
host = gentooserver
osd pool default size = 3
osd pool default min size = 2

[mds.mds0]
host = gentooserver


ceph -s times out:

gentooserver ~ # ceph -s
2018-04-06 15:58:29.861647 7f5e7f891700  0 monclient(hunting): authenticate
timed out after 300
2018-04-06 15:58:29.861672 7f5e7f891700  0 librados: client.admin
authentication error (110) Connection timed out
[errno 110] error connecting to the cluster


The cluster was working before the IP address changed. HELP
___
ceph-users mailing list
ceph-users@lists.ceph.com

[ceph-users] PG numbers don't add up?

2018-03-13 Thread Nathan Dehnel
I try to add a data pool:

OSD_STAT USED   AVAIL TOTAL HB_PEERSPG_SUM PRIMARY_PG_SUM
9 1076M  930G  931G [0,1,2,3,4,5,6,7,8]128  5
8 1076M  930G  931G [0,1,2,3,4,5,6,7,9]128 14
7 1076M  930G  931G [0,1,2,3,4,5,6,8,9]128 14
6 1076M  930G  931G [0,1,2,3,4,5,7,8,9]128 19
5 1076M  930G  931G [0,1,2,3,4,6,7,8,9]128 15
4 1076M  930G  931G [0,1,2,3,5,6,7,8,9]128 17
0 1076M  930G  931G [1,2,3,4,5,6,7,8,9]128 16
1 1076M  930G  931G [0,2,3,4,5,6,7,8,9]128  8
2 1076M  930G  931G [0,1,3,4,5,6,7,8,9]128  8
3 1076M  930G  931G [0,1,2,4,5,6,7,8,9]128 12
sum  10765M 9304G 9315G

I try to add a metadata pool:

sum 0 0 0 0 0 0 0 0
OSD_STAT USED   AVAIL TOTAL HB_PEERSPG_SUM PRIMARY_PG_SUM
9 1076M  930G  931G [0,1,2,3,4,5,6,7,8] 73 73
8 1076M  930G  931G [0,1,2,3,4,5,6,7,9] 40 40
7 1076M  930G  931G [0,1,2,3,4,5,6,8,9] 56 56
6 1076M  930G  931G [0,1,2,3,4,5,7,8,9] 42 42
5 1076M  930G  931G [0,1,2,3,4,6,7,8,9] 54 54
4 1076M  930G  931G [0,1,2,3,5,6,7,8,9] 59 59
0 1076M  930G  931G [1,2,3,4,5,6,7,8,9] 38 38
1 1076M  930G  931G [0,2,3,4,5,6,7,8,9] 57 57
2 1076M  930G  931G [0,1,3,4,5,6,7,8,9] 45 45
3 1076M  930G  931G [0,1,2,4,5,6,7,8,9] 48 48
sum  10766M 9304G 9315G

I try to add both pools:
Error ERANGE:  pg_num 128 size 10 would mean 2816 total pgs, which exceeds
max 2000 (mon_max_pg_per_osd 200 * num_in_osds 10)

That's over a thousand more PGs than both pools combined. Where are they
coming from?
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] (no subject)

2018-03-10 Thread Nathan Dehnel
Trying to create an OSD:

gentooserver ~ # ceph-volume lvm create --data /dev/sdb
Running command: ceph-authtool --gen-print-key
Running command: ceph --cluster ceph --name client.bootstrap-osd --keyring
/var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new
e70500fe-0d51-48c3-a607-414957886726
Running command: vgcreate --force --yes
ceph-a736559a-92d1-483e-9289-d2c7feed510f /dev/sdb
 stdout: Volume group "ceph-a736559a-92d1-483e-9289-d2c7feed510f"
successfully created
Running command: lvcreate --yes -l 100%FREE -n
osd-block-e70500fe-0d51-48c3-a607-414957886726
ceph-a736559a-92d1-483e-9289-d2c7feed510f
 stdout: Logical volume "osd-block-e70500fe-0d51-48c3-a607-414957886726"
created.
Running command: ceph-authtool --gen-print-key
Running command: mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-0
Running command: chown -R ceph:ceph /dev/dm-0
Running command: ln -s
/dev/ceph-a736559a-92d1-483e-9289-d2c7feed510f/osd-block-e70500fe-0d51-48c3-a607-414957886726
/var/lib/ceph/osd/ceph-0/block
Running command: ceph --cluster ceph --name client.bootstrap-osd --keyring
/var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o
/var/lib/ceph/osd/ceph-0/activate.monmap
 stderr: got monmap epoch 1
Running command: ceph-authtool /var/lib/ceph/osd/ceph-0/keyring
--create-keyring --name osd.0 --add-key
AQBEZqRalRoRCBAA03R6VshykLcZjMgQnFKDtg==
 stdout: creating /var/lib/ceph/osd/ceph-0/keyring
 stdout: added entity osd.0 auth auth(auid = 18446744073709551615
key=AQBEZqRalRoRCBAA03R6VshykLcZjMgQnFKDtg== with 0 caps)
Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/keyring
Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/
Running command: ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs
-i 0 --monmap /var/lib/ceph/osd/ceph-0/activate.monmap --keyfile -
--osd-data /var/lib/ceph/osd/ceph-0/ --osd-uuid
e70500fe-0d51-48c3-a607-414957886726 --setuser ceph --setgroup ceph
--> ceph-volume lvm prepare successful for: /dev/sdb
Running command: ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev
/dev/ceph-a736559a-92d1-483e-9289-d2c7feed510f/osd-block-e70500fe-0d51-48c3-a607-414957886726
--path /var/lib/ceph/osd/ceph-0
Running command: ln -snf
/dev/ceph-a736559a-92d1-483e-9289-d2c7feed510f/osd-block-e70500fe-0d51-48c3-a607-414957886726
/var/lib/ceph/osd/ceph-0/block
Running command: chown -R ceph:ceph /dev/dm-0
Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-0
Running command: systemctl enable
ceph-volume@lvm-0-e70500fe-0d51-48c3-a607-414957886726
 stderr: Created symlink
/etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-0-e70500fe-0d51-48c3-a607-414957886726.service
→ /lib/systemd/system/ceph-volume@.service.
Running command: systemctl start ceph-osd@0
 stderr: Job for ceph-osd@0.service failed because the control process
exited with error code.
See "systemctl status ceph-osd@0.service" and "journalctl -xe" for details.
--> Was unable to complete a new OSD, will rollback changes
--> OSD will be fully purged from the cluster, because the ID was generated
Running command: ceph osd purge osd.0 --yes-i-really-mean-it
 stderr: purged osd.0
-->  RuntimeError: command returned non-zero exit status: 1

journalctl -xe
-- Unit ceph-osd@0.service has begun starting up.
Mar 10 17:14:34 gentooserver systemd[3977]: ceph-osd@0.service: Failed to
execute command: No such file or directory
Mar 10 17:14:34 gentooserver systemd[3977]: ceph-osd@0.service: Failed at
step EXEC spawning /usr/lib/ceph/ceph-osd-prestart.sh: No such file or
directory
-- Subject: Process /usr/lib/ceph/ceph-osd-prestart.sh could not be executed
-- Defined-By: systemd
-- Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
-- 
-- The process /usr/lib/ceph/ceph-osd-prestart.sh could not be executed and
failed.
-- 
-- The error number returned by this process is 2.
Mar 10 17:14:34 gentooserver systemd[1]: ceph-osd@0.service: Control
process exited, code=exited status=203
Mar 10 17:14:34 gentooserver systemd[1]: ceph-osd@0.service: Failed with
result 'exit-code'.
Mar 10 17:14:34 gentooserver systemd[1]: Failed to start Ceph object
storage daemon osd.0.
-- Subject: Unit ceph-osd@0.service has failed
-- Defined-By: systemd
-- Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel
-- 
-- Unit ceph-osd@0.service has failed.
-- 
-- The result is RESULT.

Why is this file missing? Should I file a bug with my distro's packager?
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Fwd: ceph-volume raw disks

2018-01-25 Thread Nathan Dehnel
-- Forwarded message --
From: Nathan Dehnel <ncdeh...@gmail.com>
Date: Thu, Jan 25, 2018 at 9:49 PM
Subject: Re: [ceph-users] ceph-volume raw disks
To: David Turner <drakonst...@gmail.com>


>Did you wipe all of the existing partitions and such first?
I tried it both before and after creating an lvm physical partition.

>Which version of ceph?
12.2.1

gentooserver ~ # ceph-volume lvm zap /dev/sdb
usage: ceph-volume lvm [-h]
ceph-volume lvm: error: unrecognized arguments: zap /dev/sdb
???

On Thu, Jan 25, 2018 at 9:42 PM, David Turner <drakonst...@gmail.com> wrote:

> Did you wipe all of the existing partitions and such first?  Which version
> of ceph?  The below commands are what I ran to re-add my osds as bluestore
> after moving all data off of them.
>
> ceph-volume lvm zap /dev/sdb
> ceph-volume lvm create --bluestore --data /dev/sdb
>
> On Thu, Jan 25, 2018 at 9:41 PM Nathan Dehnel <ncdeh...@gmail.com> wrote:
>
>> The doc at http://docs.ceph.com/docs/master/ceph-volume/lvm/prepare/
>> #ceph-volume-lvm-prepare says I can pass a physical device to
>> ceph-volume. But when I try to do that:
>>
>> gentooserver ~ # ceph-volume lvm create --bluestore --data /dev/sdb
>> usage: ceph-volume lvm create [-h] [--journal JOURNAL] --data DATA
>>   [--journal-size GB] [--bluestore]
>> [--filestore]
>>   [--osd-id OSD_ID] [--osd-fsid OSD_FSID]
>> ceph-volume lvm create: error: Logical volume must be specified as
>> 'volume_group/logical_volume' but got: /dev/sdb
>>
>> Am I missing something?
>> ___
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] ceph-volume raw disks

2018-01-25 Thread Nathan Dehnel
The doc at
http://docs.ceph.com/docs/master/ceph-volume/lvm/prepare/#ceph-volume-lvm-prepare
says I can pass a physical device to ceph-volume. But when I try to do that:

gentooserver ~ # ceph-volume lvm create --bluestore --data /dev/sdb
usage: ceph-volume lvm create [-h] [--journal JOURNAL] --data DATA
  [--journal-size GB] [--bluestore]
[--filestore]
  [--osd-id OSD_ID] [--osd-fsid OSD_FSID]
ceph-volume lvm create: error: Logical volume must be specified as
'volume_group/logical_volume' but got: /dev/sdb

Am I missing something?
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] ceph command hangs

2018-01-18 Thread Nathan Dehnel
11:11.984458 7f9579ffb700  1 --
[2001:1c:d64b:91c5:3a84:dfce:8546:9982]:0/285566692 -->
[2001:1c:d64b:91c5:3a84:dfce:8546:9982]:6789/0 -- mon_command({"prefix":
"get_command_descriptions"} v 0) v1 -- 0x7f955c0041d0 con 0
2018-01-18 21:11:11.984678 7f9579ffb700  1 --
[2001:1c:d64b:91c5:3a84:dfce:8546:9982]:0/285566692 <== mon.0
[2001:1c:d64b:91c5:3a84:dfce:8546:9982]:6789/0 2  mgrmap(e 1) v1 
103+0+0 (706778617 0 0) 0x7f956c000fd0 con 0x7f9560007290
^CError EINTR: problem getting command descriptions from mon.
2018-01-18 21:11:41.999310 7f9571ffb700  1 --
[2001:1c:d64b:91c5:3a84:dfce:8546:9982]:0/285566692 >>
[2001:1c:d64b:91c5:3a84:dfce:8546:9982]:6789/0 conn(0x7f9560007290 :-1
s=STATE_OPEN pgs=5 cs=1 l=1).mark_down
2018-01-18 21:11:41.999382 7f9571ffb700  1 --
[2001:1c:d64b:91c5:3a84:dfce:8546:9982]:0/285566692 shutdown_connections
2018-01-18 21:11:41.999464 7f9571ffb700  1 --
[2001:1c:d64b:91c5:3a84:dfce:8546:9982]:0/285566692 shutdown_connections
2018-01-18 21:11:41.999488 7f9571ffb700  1 --
[2001:1c:d64b:91c5:3a84:dfce:8546:9982]:0/285566692 wait complete.
2018-01-18 21:11:41.999494 7f9571ffb700  1 --
[2001:1c:d64b:91c5:3a84:dfce:8546:9982]:0/285566692 >>
[2001:1c:d64b:91c5:3a84:dfce:8546:9982]:0/285566692 conn(0x7f957c118c90 :-1
s=STATE_NONE pgs=0 cs=0 l=0).mark_down

To my knowledge there are no firewall settings that would cause problems.
I'm running ceph -s on the monitor node.

On Tue, Jan 16, 2018 at 8:05 PM, Jean-Charles Lopez <jelo...@redhat.com>
wrote:

> Hi Nathan,
>
> I would have place the mon_host parameter and assigned it the IP address
> for your monitor host in the global section so that the client (ceph -s
> command).
>
> Have you also checked your firewall setup on your MON box?
>
> To help diagnose you can also use ceph -s --debug-ms=1 so you can follow
> the network exchange between your client machine and the MON.
>
> Regards
> JC
>
>
> On Jan 16, 2018, at 13:24, Nathan Dehnel <ncdeh...@gmail.com> wrote:
>
> I'm doing a manual setup following http://docs.ceph.
> com/docs/master/install/manual-deployment/
>
> The ceph command hangs until I kill it. I have 1 monitor service started.
> ==
> gentooserver ~ # ceph -s
> ^CError EINTR: problem getting command descriptions from mon.
> =
> gentooserver ~ # emerge -pqv ceph
> [ebuild   R   ] sys-cluster/ceph-12.2.1  USE="mgr radosgw ssl systemd
> tcmalloc -babeltrace -cephfs -fuse -jemalloc -ldap -lttng -nss -static-libs
> {-test} -xfs -zfs" CPU_FLAGS_X86="sse sse2 sse3 sse4_1 sse4_2 ssse3"
> PYTHON_TARGETS="python2_7 python3_5 -python3_4 -python3_6"
> ==
> gentooserver ~ # cat /etc/ceph/ceph.conf
> [global]
> cluster = ceph
> fsid = a736559a-92d1-483e-9289-d2c7feed510f
> ms bind ipv6 = true
> #public network = 2001:1c:d64b:91c5:3a84:dfce:8546:9982/0
> auth cluster required = none
> auth service required = none
> auth client required = none
>
> [mon]
> mon initial members = mon0
> mon host = gentooserver
> mon addr = [2001:1c:d64b:91c5:3a84:dfce:8546:9982]:6789
>
>
> [mon.mon0]
> host = gentooserver
> mon addr = [2001:1c:d64b:91c5:3a84:dfce:8546:9982]:6789
>
> [osd]
> osd journal size = 1
> osd crush chooseleaf type = 0
> ==
> gentooserver ~ # monmaptool --print /tmp/monmap
> monmaptool: monmap file /tmp/monmap
> epoch 0
> fsid a736559a-92d1-483e-9289-d2c7feed510f
> last_changed 2018-01-14 16:50:59.838277
> created 2018-01-14 16:50:59.838277
> 0: [2001:1c:d64b:91c5:3a84:dfce:8546:9982]:6789/0 mon.mon0
> ==
> gentooserver ~ # systemctl status ceph-mon@mon0 | cat
> ● ceph-mon@mon0.service - Ceph cluster monitor daemon
>Loaded: loaded (/usr/lib/systemd/system/ceph-mon@.service; enabled;
> vendor preset: disabled)
>Active: active (running) since Tue 2018-01-16 14:50:18 CST; 17min ago
>  Main PID: 75938 (ceph-mon)
>CGroup: /system.slice/system-ceph\x2dmon.slice/ceph-mon@mon0.service
>└─75938 /usr/bin/ceph-mon -f --cluster ceph --id mon0 --setuser
> ceph --setgroup ceph
>
> Jan 16 14:50:18 gentooserver systemd[1]: Started Ceph cluster monitor
> daemon.
> Jan 16 14:50:18 gentooserver ceph-mon[75938]: 2018-01-16 14:50:18.977494
> 7ff07d4cef80 -1 distro_detect - can't detect distro_version
> ===
> gentooserver ~ # cat /var/log/ceph/ceph.log
> 2018-01-16 14:50:18.977541 mon.mon0 mon.0 
> [2001:1c:d64b:91c5:3a84:dfce:8546:9982]:6789/0
> 1 : cluster [INF] mon.mon0@0 won leader election with quorum 0
> 2018-01

[ceph-users] ceph command hangs

2018-01-16 Thread Nathan Dehnel
I'm doing a manual setup following
http://docs.ceph.com/docs/master/install/manual-deployment/

The ceph command hangs until I kill it. I have 1 monitor service started.
==
gentooserver ~ # ceph -s
^CError EINTR: problem getting command descriptions from mon.
=
gentooserver ~ # emerge -pqv ceph
[ebuild   R   ] sys-cluster/ceph-12.2.1  USE="mgr radosgw ssl systemd
tcmalloc -babeltrace -cephfs -fuse -jemalloc -ldap -lttng -nss -static-libs
{-test} -xfs -zfs" CPU_FLAGS_X86="sse sse2 sse3 sse4_1 sse4_2 ssse3"
PYTHON_TARGETS="python2_7 python3_5 -python3_4 -python3_6"
==
gentooserver ~ # cat /etc/ceph/ceph.conf
[global]
cluster = ceph
fsid = a736559a-92d1-483e-9289-d2c7feed510f
ms bind ipv6 = true
#public network = 2001:1c:d64b:91c5:3a84:dfce:8546:9982/0
auth cluster required = none
auth service required = none
auth client required = none

[mon]
mon initial members = mon0
mon host = gentooserver
mon addr = [2001:1c:d64b:91c5:3a84:dfce:8546:9982]:6789


[mon.mon0]
host = gentooserver
mon addr = [2001:1c:d64b:91c5:3a84:dfce:8546:9982]:6789

[osd]
osd journal size = 1
osd crush chooseleaf type = 0
==
gentooserver ~ # monmaptool --print /tmp/monmap
monmaptool: monmap file /tmp/monmap
epoch 0
fsid a736559a-92d1-483e-9289-d2c7feed510f
last_changed 2018-01-14 16:50:59.838277
created 2018-01-14 16:50:59.838277
0: [2001:1c:d64b:91c5:3a84:dfce:8546:9982]:6789/0 mon.mon0
==
gentooserver ~ # systemctl status ceph-mon@mon0 | cat
● ceph-mon@mon0.service - Ceph cluster monitor daemon
   Loaded: loaded (/usr/lib/systemd/system/ceph-mon@.service; enabled;
vendor preset: disabled)
   Active: active (running) since Tue 2018-01-16 14:50:18 CST; 17min ago
 Main PID: 75938 (ceph-mon)
   CGroup: /system.slice/system-ceph\x2dmon.slice/ceph-mon@mon0.service
   └─75938 /usr/bin/ceph-mon -f --cluster ceph --id mon0 --setuser
ceph --setgroup ceph

Jan 16 14:50:18 gentooserver systemd[1]: Started Ceph cluster monitor
daemon.
Jan 16 14:50:18 gentooserver ceph-mon[75938]: 2018-01-16 14:50:18.977494
7ff07d4cef80 -1 distro_detect - can't detect distro_version
===
gentooserver ~ # cat /var/log/ceph/ceph.log
2018-01-16 14:50:18.977541 mon.mon0 mon.0
[2001:1c:d64b:91c5:3a84:dfce:8546:9982]:6789/0 1 : cluster [INF] mon.mon0@0
won leader election with quorum 0
2018-01-16 14:50:18.977656 mon.mon0 mon.0
[2001:1c:d64b:91c5:3a84:dfce:8546:9982]:6789/0 2 : cluster [INF] monmap e1:
1 mons at {mon0=[2001:1c:d64b:91c5:3a84:dfce:8546:9982]:6789/0}
---
gentooserver ~ # cat /var/log/ceph/ceph-mon.mon0.log
2018-01-16 14:50:18.760533 7ff07d4cef80  0 set uid:gid to 110:239
(ceph:ceph)
2018-01-16 14:50:18.760549 7ff07d4cef80  0 ceph version 12.2.1
(3e7492b9ada8bdc9a5cd0feafd42fbca27f9c38e) luminous (stable), process
(unknown), pid 75938
2018-01-16 14:50:18.760591 7ff07d4cef80  0 pidfile_write: ignore empty
--pid-file
2018-01-16 14:50:18.765642 7ff07d4cef80  0 load: jerasure load: lrc load:
isa
2018-01-16 14:50:18.765702 7ff07d4cef80  0  set rocksdb option compression
= kNoCompression
2018-01-16 14:50:18.765709 7ff07d4cef80  0  set rocksdb option
write_buffer_size = 33554432
2018-01-16 14:50:18.765722 7ff07d4cef80  0  set rocksdb option compression
= kNoCompression
2018-01-16 14:50:18.765726 7ff07d4cef80  0  set rocksdb option
write_buffer_size = 33554432
2018-01-16 14:50:18.765798 7ff07d4cef80  4 rocksdb: RocksDB version: 5.4.0

2018-01-16 14:50:18.765804 7ff07d4cef80  4 rocksdb: Git sha
rocksdb_build_git_sha:@0@
2018-01-16 14:50:18.765806 7ff07d4cef80  4 rocksdb: Compile date Dec 14 2017
2018-01-16 14:50:18.765808 7ff07d4cef80  4 rocksdb: DB SUMMARY

2018-01-16 14:50:18.765837 7ff07d4cef80  4 rocksdb: CURRENT file:  CURRENT

2018-01-16 14:50:18.765840 7ff07d4cef80  4 rocksdb: IDENTITY file:  IDENTITY

2018-01-16 14:50:18.765843 7ff07d4cef80  4 rocksdb: MANIFEST file:
MANIFEST-11 size: 210 Bytes

2018-01-16 14:50:18.765845 7ff07d4cef80  4 rocksdb: SST files in
/var/lib/ceph/mon/ceph-mon0/store.db dir, Total Num: 3, files: 04.sst
07.sst 10.sst

2018-01-16 14:50:18.765846 7ff07d4cef80  4 rocksdb: Write Ahead Log file in
/var/lib/ceph/mon/ceph-mon0/store.db: 12.log size: 9161 ;

2018-01-16 14:50:18.765847 7ff07d4cef80  4 rocksdb:
 Options.error_if_exists: 0
2018-01-16 14:50:18.765848 7ff07d4cef80  4 rocksdb:
 Options.create_if_missing: 0
2018-01-16 14:50:18.765848 7ff07d4cef80  4 rocksdb:
 Options.paranoid_checks: 1
2018-01-16 14:50:18.765849 7ff07d4cef80  4 rocksdb:
 Options.env: 0x55bc0ff83a60
2018-01-16 14:50:18.765849 7ff07d4cef80  4 rocksdb:
Options.info_log: 0x55bc10a53ae0
2018-01-16 14:50:18.765850 7ff07d4cef80  4 rocksdb:
  

Re: [ceph-users] permission denied, unable to bind socket

2018-01-07 Thread Nathan Dehnel
Ok I fixed the address error. The service is able to start now. ceph -s
hangs though.

gentooserver ~ # ceph -s
^CError EINTR: problem getting command descriptions from mon.

I'm not sure how to fix the permissions issue. /var/run/ceph is a temporary
directory so I can't just chown it.

On Sat, Jan 6, 2018 at 7:40 PM, Nathan Dehnel <ncdeh...@gmail.com> wrote:

> So I'm following the guide at http://docs.ceph.com/docs/
> master/install/manual-deployment/
>
> ceph-mon@gentooserver.service fails.
>
> Jan 06 19:12:40 gentooserver systemd[1]: Started Ceph cluster monitor
> daemon.
> Jan 06 19:12:41 gentooserver ceph-mon[2674]: warning: unable to create
> /var/run/ceph: (13) Permission denied
> Jan 06 19:12:41 gentooserver ceph-mon[2674]: 2018-01-06 19:12:41.000507
> 7f3163008f80 -1 asok(0x563d0f97d2c0) AdminSocketConfigObs::init: failed:
> AdminSocket::bind_and_listen: failed to bind the UNIX domain socket to
> '/var/run/ceph/ceph-mon.gentooserver.asok': (2) No such file or directory
> Jan 06 19:12:41 gentooserver ceph-mon[2674]: 2018-01-06 19:12:41.152781
> 7f3163008f80 -1  Processor -- bind unable to bind to
> [2001:1c:d64b:91c5:3a84:dfce:8546:998]:6789/0: (99) Cannot assign
> requested address
> Jan 06 19:12:41 gentooserver ceph-mon[2674]: 2018-01-06 19:12:41.152789
> 7f3163008f80 -1  Processor -- bind was unable to bind. Trying again in 5
> seconds
> Jan 06 19:12:46 gentooserver ceph-mon[2674]: 2018-01-06 19:12:46.152982
> 7f3163008f80 -1  Processor -- bind unable to bind to
> [2001:1c:d64b:91c5:3a84:dfce:8546:998]:6789/0: (99) Cannot assign
> requested address
> Jan 06 19:12:46 gentooserver ceph-mon[2674]: 2018-01-06 19:12:46.152996
> 7f3163008f80 -1  Processor -- bind was unable to bind. Trying again in 5
> seconds
> Jan 06 19:12:51 gentooserver ceph-mon[2674]: 2018-01-06 19:12:51.153218
> 7f3163008f80 -1  Processor -- bind unable to bind to
> [2001:1c:d64b:91c5:3a84:dfce:8546:998]:6789/0: (99) Cannot assign
> requested address
> Jan 06 19:12:51 gentooserver ceph-mon[2674]: 2018-01-06 19:12:51.153244
> 7f3163008f80 -1  Processor -- bind was unable to bind after 3 attempts:
> (99) Cannot assign requested address
> Jan 06 19:12:51 gentooserver ceph-mon[2674]: 2018-01-06 19:12:51.153250
> 7f3163008f80 -1 unable to bind monitor to [2001:1c:d64b:91c5:3a84:dfce:
> 8546:998]:6789/0
> Jan 06 19:12:51 gentooserver systemd[1]: ceph-mon@gentooserver.service:
> Main process exited, code=exited, status=1/FAILURE
> Jan 06 19:12:51 gentooserver systemd[1]: ceph-mon@gentooserver.service:
> Unit entered failed state.
> Jan 06 19:12:51 gentooserver systemd[1]: ceph-mon@gentooserver.service:
> Failed with result 'exit-code'.
> Jan 06 19:13:01 gentooserver systemd[1]: ceph-mon@gentooserver.service:
> Service hold-off time over, scheduling restart.
> Jan 06 19:13:01 gentooserver systemd[1]: Stopped Ceph cluster monitor
> daemon.
> Jan 06 19:13:01 gentooserver systemd[1]: ceph-mon@gentooserver.service:
> Start request repeated too quickly.
> Jan 06 19:13:01 gentooserver systemd[1]: Failed to start Ceph cluster
> monitor daemon.
> Jan 06 19:13:01 gentooserver systemd[1]: ceph-mon@gentooserver.service:
> Unit entered failed state.
> Jan 06 19:13:01 gentooserver systemd[1]: ceph-mon@gentooserver.service:
> Failed with result 'exit-code'.
>
> cat /etc/ceph/ceph.conf
> [global]
> fsid = a736559a-92d1-483e-9289-d2c7feed510f
> ms bind ipv6 = true
> mon initial members = gentooserver
>
> [mon.mona]
> host = gentooserver
> mon addr = [2001:1c:d64b:91c5:3a84:dfce:8546:9982]
>
> [osd]
> osd journal size = 1
>
> I'm not sure if the problem is the permissions error, or the IP address
> appearing to get truncated in the output, or both.
>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] permission denied, unable to bind socket

2018-01-06 Thread Nathan Dehnel
So I'm following the guide at
http://docs.ceph.com/docs/master/install/manual-deployment/

ceph-mon@gentooserver.service fails.

Jan 06 19:12:40 gentooserver systemd[1]: Started Ceph cluster monitor
daemon.
Jan 06 19:12:41 gentooserver ceph-mon[2674]: warning: unable to create
/var/run/ceph: (13) Permission denied
Jan 06 19:12:41 gentooserver ceph-mon[2674]: 2018-01-06 19:12:41.000507
7f3163008f80 -1 asok(0x563d0f97d2c0) AdminSocketConfigObs::init: failed:
AdminSocket::bind_and_listen: failed to bind the UNIX domain socket to
'/var/run/ceph/ceph-mon.gentooserver.asok': (2) No such file or directory
Jan 06 19:12:41 gentooserver ceph-mon[2674]: 2018-01-06 19:12:41.152781
7f3163008f80 -1  Processor -- bind unable to bind to
[2001:1c:d64b:91c5:3a84:dfce:8546:998]:6789/0: (99) Cannot assign requested
address
Jan 06 19:12:41 gentooserver ceph-mon[2674]: 2018-01-06 19:12:41.152789
7f3163008f80 -1  Processor -- bind was unable to bind. Trying again in 5
seconds
Jan 06 19:12:46 gentooserver ceph-mon[2674]: 2018-01-06 19:12:46.152982
7f3163008f80 -1  Processor -- bind unable to bind to
[2001:1c:d64b:91c5:3a84:dfce:8546:998]:6789/0: (99) Cannot assign requested
address
Jan 06 19:12:46 gentooserver ceph-mon[2674]: 2018-01-06 19:12:46.152996
7f3163008f80 -1  Processor -- bind was unable to bind. Trying again in 5
seconds
Jan 06 19:12:51 gentooserver ceph-mon[2674]: 2018-01-06 19:12:51.153218
7f3163008f80 -1  Processor -- bind unable to bind to
[2001:1c:d64b:91c5:3a84:dfce:8546:998]:6789/0: (99) Cannot assign requested
address
Jan 06 19:12:51 gentooserver ceph-mon[2674]: 2018-01-06 19:12:51.153244
7f3163008f80 -1  Processor -- bind was unable to bind after 3 attempts:
(99) Cannot assign requested address
Jan 06 19:12:51 gentooserver ceph-mon[2674]: 2018-01-06 19:12:51.153250
7f3163008f80 -1 unable to bind monitor to
[2001:1c:d64b:91c5:3a84:dfce:8546:998]:6789/0
Jan 06 19:12:51 gentooserver systemd[1]: ceph-mon@gentooserver.service:
Main process exited, code=exited, status=1/FAILURE
Jan 06 19:12:51 gentooserver systemd[1]: ceph-mon@gentooserver.service:
Unit entered failed state.
Jan 06 19:12:51 gentooserver systemd[1]: ceph-mon@gentooserver.service:
Failed with result 'exit-code'.
Jan 06 19:13:01 gentooserver systemd[1]: ceph-mon@gentooserver.service:
Service hold-off time over, scheduling restart.
Jan 06 19:13:01 gentooserver systemd[1]: Stopped Ceph cluster monitor
daemon.
Jan 06 19:13:01 gentooserver systemd[1]: ceph-mon@gentooserver.service:
Start request repeated too quickly.
Jan 06 19:13:01 gentooserver systemd[1]: Failed to start Ceph cluster
monitor daemon.
Jan 06 19:13:01 gentooserver systemd[1]: ceph-mon@gentooserver.service:
Unit entered failed state.
Jan 06 19:13:01 gentooserver systemd[1]: ceph-mon@gentooserver.service:
Failed with result 'exit-code'.

cat /etc/ceph/ceph.conf
[global]
fsid = a736559a-92d1-483e-9289-d2c7feed510f
ms bind ipv6 = true
mon initial members = gentooserver

[mon.mona]
host = gentooserver
mon addr = [2001:1c:d64b:91c5:3a84:dfce:8546:9982]

[osd]
osd journal size = 1

I'm not sure if the problem is the permissions error, or the IP address
appearing to get truncated in the output, or both.
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] ceph.conf not found

2018-01-04 Thread Nathan Dehnel
Hey, I get this error:

gentooserver ~ # ceph -s
2018-01-04 14:38:35.390154 7f0a6bae8700 -1 Errors while parsing config file!
2018-01-04 14:38:35.390157 7f0a6bae8700 -1 parse_file: cannot open
/etc/ceph/ceph.conf: (2) No such file or directory
2018-01-04 14:38:35.390158 7f0a6bae8700 -1 parse_file: cannot open
~/.ceph/ceph.conf: (2) No such file or directory
2018-01-04 14:38:35.390158 7f0a6bae8700 -1 parse_file: cannot open
ceph.conf: (2) No such file or directory
Error initializing cluster client: ObjectNotFound('error calling
conf_read_file',)

I don't have a ceph.conf because my cluster name is "home" and the guide at
http://docs.ceph.com/docs/master/install/manual-deployment/ says to name
the configuration file after the cluster name. What should I do to resolve
this?
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com