Re: [openstack-dev] [kolla] ceph osd deploy fails

2018-09-26 Thread Eduardo Gonzalez
Hi, what version of rocky are you using. Maybe was in the middle of a
backport which temporally broke ceph.

Could you try latest stable/rocky branch?

It is now working properly.

Regards

On Wed, Sep 26, 2018, 2:32 PM Florian Engelmann <
florian.engelm...@everyware.ch> wrote:

> Hi,
>
> I tried to deploy Rocky in a multinode setup but ceph-osd fails with:
>
>
> failed: [xxx-poc2] (item=[0, {u'fs_uuid': u'', u'bs_wal_label':
> u'', u'external_journal': False, u'bs_blk_label': u'',
> u'bs_db_partition_num': u'', u'journal_device': u'', u'journal': u'',
> u'partition': u'/dev/nvme0n1', u'bs_wal_partition_num': u'',
> u'fs_label': u'', u'journal_num': 0, u'bs_wal_device': u'',
> u'partition_num': u'1', u'bs_db_label': u'', u'bs_blk_partition_num':
> u'', u'device': u'/dev/nvme0n1', u'bs_db_device': u'',
> u'partition_label': u'KOLLA_CEPH_OSD_BOOTSTRAP_BS', u'bs_blk_device':
> u''}]) => {
>  "changed": true,
>  "item": [
>  0,
>  {
>  "bs_blk_device": "",
>  "bs_blk_label": "",
>  "bs_blk_partition_num": "",
>  "bs_db_device": "",
>  "bs_db_label": "",
>  "bs_db_partition_num": "",
>  "bs_wal_device": "",
>  "bs_wal_label": "",
>  "bs_wal_partition_num": "",
>  "device": "/dev/nvme0n1",
>  "external_journal": false,
>  "fs_label": "",
>  "fs_uuid": "",
>  "journal": "",
>  "journal_device": "",
>  "journal_num": 0,
>  "partition": "/dev/nvme0n1",
>  "partition_label": "KOLLA_CEPH_OSD_BOOTSTRAP_BS",
>  "partition_num": "1"
>  }
>  ]
> }
>
> MSG:
>
> Container exited with non-zero return code 2
>
> We tried to debug the error message by starting the container with a
> modified endpoint but we are stuck at the following point right now:
>
>
> docker run  -e "HOSTNAME=10.0.153.11" -e "JOURNAL_DEV=" -e
> "JOURNAL_PARTITION=" -e "JOURNAL_PARTITION_NUM=0" -e
> "KOLLA_BOOTSTRAP=null" -e "KOLLA_CONFIG_STRATEGY=COPY_ALWAYS" -e
> "KOLLA_SERVICE_NAME=bootstrap-osd-0" -e "OSD_BS_BLK_DEV=" -e
> "OSD_BS_BLK_LABEL=" -e "OSD_BS_BLK_PARTNUM=" -e "OSD_BS_DB_DEV=" -e
> "OSD_BS_DB_LABEL=" -e "OSD_BS_DB_PARTNUM=" -e "OSD_BS_DEV=/dev/nvme0n1"
> -e "OSD_BS_LABEL=KOLLA_CEPH_OSD_BOOTSTRAP_BS" -e "OSD_BS_PARTNUM=1" -e
> "OSD_BS_WAL_DEV=" -e "OSD_BS_WAL_LABEL=" -e "OSD_BS_WAL_PARTNUM=" -e
> "OSD_DEV=/dev/nvme0n1" -e "OSD_FILESYSTEM=xfs" -e "OSD_INITIAL_WEIGHT=1"
> -e "OSD_PARTITION=/dev/nvme0n1" -e "OSD_PARTITION_NUM=1" -e
> "OSD_STORETYPE=bluestore" -e "USE_EXTERNAL_JOURNAL=false"   -v
> "/etc/kolla//ceph-osd/:/var/lib/kolla/config_files/:ro" -v
> "/etc/localtime:/etc/localtime:ro" -v "/dev/:/dev/" -v
> "kolla_logs:/var/log/kolla/" -ti --privileged=true --entrypoint
> /bin/bash
>
> 10.0.128.7:5000/openstack/openstack-kolla-cfg/ubuntu-source-ceph-osd:7.0.0.3
>
>
>
> cat /var/lib/kolla/config_files/ceph.client.admin.keyring >
> /etc/ceph/ceph.client.admin.keyring
>
>
> cat /var/lib/kolla/config_files/ceph.conf > /etc/ceph/ceph.conf
>
>
> (bootstrap-osd-0)[root@985e2dee22bc /]# /usr/bin/ceph-osd -d
> --public-addr 10.0.153.11 --cluster-addr 10.0.153.11
> usage: ceph-osd -i  [flags]
>--osd-data PATH data directory
>--osd-journal PATH
>  journal file or block device
>--mkfscreate a [new] data directory
>--mkkey   generate a new secret key. This is normally used in
> combination with --mkfs
>--convert-filestore
>  run any pending upgrade operations
>--flush-journal   flush all data out of journal
>--mkjournal   initialize a new journal
>--check-wants-journal
>  check whether a journal is desired
>--check-allows-journal
>  check whether a journal is allowed
>--check-needs-journal
>  check whether a journal is required
>--debug_osdset debug level (e.g. 10)
>--get-device-fsid PATH
>  get OSD fsid for the given block device
>
>--conf/-c FILEread configuration from the given configuration file
>--id/-i IDset ID portion of my name
>--name/-n TYPE.ID set name
>--cluster NAMEset cluster name (default: ceph)
>--setuser USERset uid to user or uid (and gid to user's gid)
>--setgroup GROUP  set gid to group or gid
>--version show version and quit
>
>-drun in foreground, log to stderr.
>-frun in foreground, log to usual location.
>--debug_ms N  set message debug level (e.g. 1)
> 2018-09-26 12:28:07.801066 7fbda64b4e40  0 ceph version 12.2.4
> (52085d5249a80c5f5121a76d6288429f35e4e77b) luminous (stable), process
> (unknown), pid 46
> 2018-09-26 12:28:07.801078 7fbda64b4e40 -1 must specify '-i #' where #
> is the osd number
>
>
> But it looks like "-i" is not set anywere?
>
> grep command
> 

[openstack-dev] [kolla] ceph osd deploy fails

2018-09-26 Thread Florian Engelmann

Hi,

I tried to deploy Rocky in a multinode setup but ceph-osd fails with:


failed: [xxx-poc2] (item=[0, {u'fs_uuid': u'', u'bs_wal_label': 
u'', u'external_journal': False, u'bs_blk_label': u'', 
u'bs_db_partition_num': u'', u'journal_device': u'', u'journal': u'', 
u'partition': u'/dev/nvme0n1', u'bs_wal_partition_num': u'', 
u'fs_label': u'', u'journal_num': 0, u'bs_wal_device': u'', 
u'partition_num': u'1', u'bs_db_label': u'', u'bs_blk_partition_num': 
u'', u'device': u'/dev/nvme0n1', u'bs_db_device': u'', 
u'partition_label': u'KOLLA_CEPH_OSD_BOOTSTRAP_BS', u'bs_blk_device': 
u''}]) => {

"changed": true,
"item": [
0,
{
"bs_blk_device": "",
"bs_blk_label": "",
"bs_blk_partition_num": "",
"bs_db_device": "",
"bs_db_label": "",
"bs_db_partition_num": "",
"bs_wal_device": "",
"bs_wal_label": "",
"bs_wal_partition_num": "",
"device": "/dev/nvme0n1",
"external_journal": false,
"fs_label": "",
"fs_uuid": "",
"journal": "",
"journal_device": "",
"journal_num": 0,
"partition": "/dev/nvme0n1",
"partition_label": "KOLLA_CEPH_OSD_BOOTSTRAP_BS",
"partition_num": "1"
}
]
}

MSG:

Container exited with non-zero return code 2

We tried to debug the error message by starting the container with a 
modified endpoint but we are stuck at the following point right now:



docker run  -e "HOSTNAME=10.0.153.11" -e "JOURNAL_DEV=" -e 
"JOURNAL_PARTITION=" -e "JOURNAL_PARTITION_NUM=0" -e 
"KOLLA_BOOTSTRAP=null" -e "KOLLA_CONFIG_STRATEGY=COPY_ALWAYS" -e 
"KOLLA_SERVICE_NAME=bootstrap-osd-0" -e "OSD_BS_BLK_DEV=" -e 
"OSD_BS_BLK_LABEL=" -e "OSD_BS_BLK_PARTNUM=" -e "OSD_BS_DB_DEV=" -e 
"OSD_BS_DB_LABEL=" -e "OSD_BS_DB_PARTNUM=" -e "OSD_BS_DEV=/dev/nvme0n1" 
-e "OSD_BS_LABEL=KOLLA_CEPH_OSD_BOOTSTRAP_BS" -e "OSD_BS_PARTNUM=1" -e 
"OSD_BS_WAL_DEV=" -e "OSD_BS_WAL_LABEL=" -e "OSD_BS_WAL_PARTNUM=" -e 
"OSD_DEV=/dev/nvme0n1" -e "OSD_FILESYSTEM=xfs" -e "OSD_INITIAL_WEIGHT=1" 
-e "OSD_PARTITION=/dev/nvme0n1" -e "OSD_PARTITION_NUM=1" -e 
"OSD_STORETYPE=bluestore" -e "USE_EXTERNAL_JOURNAL=false"   -v 
"/etc/kolla//ceph-osd/:/var/lib/kolla/config_files/:ro" -v 
"/etc/localtime:/etc/localtime:ro" -v "/dev/:/dev/" -v 
"kolla_logs:/var/log/kolla/" -ti --privileged=true --entrypoint 
/bin/bash 
10.0.128.7:5000/openstack/openstack-kolla-cfg/ubuntu-source-ceph-osd:7.0.0.3




cat /var/lib/kolla/config_files/ceph.client.admin.keyring > 
/etc/ceph/ceph.client.admin.keyring



cat /var/lib/kolla/config_files/ceph.conf > /etc/ceph/ceph.conf


(bootstrap-osd-0)[root@985e2dee22bc /]# /usr/bin/ceph-osd -d 
--public-addr 10.0.153.11 --cluster-addr 10.0.153.11

usage: ceph-osd -i  [flags]
  --osd-data PATH data directory
  --osd-journal PATH
journal file or block device
  --mkfscreate a [new] data directory
  --mkkey   generate a new secret key. This is normally used in 
combination with --mkfs

  --convert-filestore
run any pending upgrade operations
  --flush-journal   flush all data out of journal
  --mkjournal   initialize a new journal
  --check-wants-journal
check whether a journal is desired
  --check-allows-journal
check whether a journal is allowed
  --check-needs-journal
check whether a journal is required
  --debug_osdset debug level (e.g. 10)
  --get-device-fsid PATH
get OSD fsid for the given block device

  --conf/-c FILEread configuration from the given configuration file
  --id/-i IDset ID portion of my name
  --name/-n TYPE.ID set name
  --cluster NAMEset cluster name (default: ceph)
  --setuser USERset uid to user or uid (and gid to user's gid)
  --setgroup GROUP  set gid to group or gid
  --version show version and quit

  -drun in foreground, log to stderr.
  -frun in foreground, log to usual location.
  --debug_ms N  set message debug level (e.g. 1)
2018-09-26 12:28:07.801066 7fbda64b4e40  0 ceph version 12.2.4 
(52085d5249a80c5f5121a76d6288429f35e4e77b) luminous (stable), process 
(unknown), pid 46
2018-09-26 12:28:07.801078 7fbda64b4e40 -1 must specify '-i #' where # 
is the osd number



But it looks like "-i" is not set anywere?

grep command 
/opt/stack/kolla-ansible/ansible/roles/ceph/templates/ceph-osd.json.j2
"command": "/usr/bin/ceph-osd -f --public-addr {{ 
hostvars[inventory_hostname]['ansible_' + 
storage_interface]['ipv4']['address'] }} --cluster-addr {{ 
hostvars[inventory_hostname]['ansible_' + 
cluster_interface]['ipv4']['address'] }}",


What's wrong with our setup?

All the best,
Flo


--

EveryWare AG
Florian Engelmann
Systems Engineer
Zurlindenstrasse 52a
CH-8003 Zürich

tel: +41 44 466 60 00
fax: +41 44 466 60 10
mail: