Hi Eugen. Entirely my missunderstanding, I thought there would be something at boot time (what would certainly not make any sense at all). Sorry.
Before stage 3 I ran the commands you suggested on the nodes, and only one got me the output below: ########################### # grep -C5 sda4 /var/log/messages 2018-08-28T08:26:50.635077-03:00 polar kernel: [ 3.029809] ata2.00: ATAPI: PLDS DVD+/-RW DU-8A5LH, 6D1M, max UDMA/133 2018-08-28T08:26:50.635080-03:00 polar kernel: [ 3.030616] ata2.00: configured for UDMA/133 2018-08-28T08:26:50.635082-03:00 polar kernel: [ 3.038249] scsi 1:0:0:0: CD-ROM PLDS DVD+-RW DU-8A5LH 6D1M PQ: 0 ANSI: 5 2018-08-28T08:26:50.635085-03:00 polar kernel: [ 3.048102] usb 1-6: new low-speed USB device number 2 using xhci_hcd 2018-08-28T08:26:50.635095-03:00 polar kernel: [ 3.051408] scsi 1:0:0:0: Attached scsi generic sg1 type 5 2018-08-28T08:26:50.635098-03:00 polar kernel: [ 3.079763] sda: sda1 sda2 sda3 sda4 2018-08-28T08:26:50.635101-03:00 polar kernel: [ 3.080548] sd 0:0:0:0: [sda] Attached SCSI disk 2018-08-28T08:26:50.635104-03:00 polar kernel: [ 3.109021] sr 1:0:0:0: [sr0] scsi3-mmc drive: 24x/24x writer cd/rw xa/form2 cdda tray 2018-08-28T08:26:50.635106-03:00 polar kernel: [ 3.109025] cdrom: Uniform CD-ROM driver Revision: 3.20 2018-08-28T08:26:50.635109-03:00 polar kernel: [ 3.109246] sr 1:0:0:0: Attached scsi CD-ROM sr0 2018-08-28T08:26:50.635112-03:00 polar kernel: [ 3.206490] usb 1-6: New USB device found, idVendor=413c, idProduct=2113 -- 2018-08-28T10:11:10.512604-03:00 polar os-prober: debug: running /usr/lib/os-probes/mounted/83haiku on mounted /dev/sda1 2018-08-28T10:11:10.516374-03:00 polar 83haiku: debug: /dev/sda1 is not a BeFS partition: exiting 2018-08-28T10:11:10.517805-03:00 polar os-prober: debug: running /usr/lib/os-probes/mounted/90linux-distro on mounted /dev/sda1 2018-08-28T10:11:10.523382-03:00 polar os-prober: debug: running /usr/lib/os-probes/mounted/90solaris on mounted /dev/sda1 2018-08-28T10:11:10.529317-03:00 polar os-prober: debug: /dev/sda2: is active swap 2018-08-28T10:11:10.539818-03:00 polar os-prober: debug: running /usr/lib/os-probes/50mounted-tests on /dev/sda4 2018-08-28T10:11:10.669852-03:00 polar systemd-udevd[456]: Network interface NamePolicy= disabled by default. 2018-08-28T10:11:10.705602-03:00 polar systemd-udevd[456]: Specified group 'plugdev' unknown 2018-08-28T10:11:10.812270-03:00 polar 50mounted-tests: debug: mounted using GRUB xfs filesystem driver 2018-08-28T10:11:10.817141-03:00 polar 50mounted-tests: debug: running subtest /usr/lib/os-probes/mounted/05efi 2018-08-28T10:11:10.832257-03:00 polar 05efi: debug: /dev/sda4 is xfs partition: exiting 2018-08-28T10:11:10.837353-03:00 polar 50mounted-tests: debug: running subtest /usr/lib/os-probes/mounted/10freedos 2018-08-28T10:11:10.851042-03:00 polar 10freedos: debug: /dev/sda4 is not a FAT partition: exiting 2018-08-28T10:11:10.854580-03:00 polar 50mounted-tests: debug: running subtest /usr/lib/os-probes/mounted/10qnx 2018-08-28T10:11:10.863539-03:00 polar 10qnx: debug: /dev/sda4 is not a QNX4 partition: exiting 2018-08-28T10:11:10.865876-03:00 polar 50mounted-tests: debug: running subtest /usr/lib/os-probes/mounted/20macosx 2018-08-28T10:11:10.871781-03:00 polar macosx-prober: debug: /dev/sda4 is not an HFS+ partition: exiting 2018-08-28T10:11:10.873708-03:00 polar 50mounted-tests: debug: running subtest /usr/lib/os-probes/mounted/20microsoft 2018-08-28T10:11:10.879146-03:00 polar 20microsoft: debug: Skipping legacy bootloaders on UEFI system 2018-08-28T10:11:10.880798-03:00 polar 50mounted-tests: debug: running subtest /usr/lib/os-probes/mounted/30utility 2018-08-28T10:11:10.885707-03:00 polar 30utility: debug: /dev/sda4 is not a FAT partition: exiting 2018-08-28T10:11:10.887422-03:00 polar 50mounted-tests: debug: running subtest /usr/lib/os-probes/mounted/40lsb 2018-08-28T10:11:10.892547-03:00 polar 50mounted-tests: debug: running subtest /usr/lib/os-probes/mounted/70hurd 2018-08-28T10:11:10.897110-03:00 polar 50mounted-tests: debug: running subtest /usr/lib/os-probes/mounted/80minix 2018-08-28T10:11:10.901133-03:00 polar 50mounted-tests: debug: running subtest /usr/lib/os-probes/mounted/83haiku 2018-08-28T10:11:10.904998-03:00 polar 83haiku: debug: /dev/sda4 is not a BeFS partition: exiting 2018-08-28T10:11:10.906289-03:00 polar 50mounted-tests: debug: running subtest /usr/lib/os-probes/mounted/90linux-distro 2018-08-28T10:11:10.912016-03:00 polar 50mounted-tests: debug: running subtest /usr/lib/os-probes/mounted/90solaris 2018-08-28T10:11:10.915838-03:00 polar 50mounted-tests: debug: running subtest /usr/lib/os-probes/mounted/efi 2018-08-28T10:11:11.757030-03:00 polar [RPM][4789]: erase kernel-default-4.12.14-lp150.12.16.1.x86_64: success 2018-08-28T10:11:11.757912-03:00 polar [RPM][4789]: Transaction ID 5b8549e8 finished: 0 -- 2018-08-28T10:13:08.815753-03:00 polar kernel: [ 2.885213] ata2.00: configured for UDMA/133 2018-08-28T10:13:08.815755-03:00 polar kernel: [ 2.885797] ata3: SATA link down (SStatus 4 SControl 300) 2018-08-28T10:13:08.815758-03:00 polar kernel: [ 2.892901] scsi 1:0:0:0: CD-ROM PLDS DVD+-RW DU-8A5LH 6D1M PQ: 0 ANSI: 5 2018-08-28T10:13:08.815762-03:00 polar kernel: [ 2.904256] usb 1-6: new low-speed USB device number 2 using xhci_hcd 2018-08-28T10:13:08.815764-03:00 polar kernel: [ 2.911580] scsi 1:0:0:0: Attached scsi generic sg1 type 5 2018-08-28T10:13:08.815777-03:00 polar kernel: [ 2.950697] sda: sda1 sda2 sda3 sda4 2018-08-28T10:13:08.815780-03:00 polar kernel: [ 2.952410] sd 0:0:0:0: [sda] Attached SCSI disk 2018-08-28T10:13:08.815784-03:00 polar kernel: [ 2.985020] sr 1:0:0:0: [sr0] scsi3-mmc drive: 24x/24x writer cd/rw xa/form2 cdda tray 2018-08-28T10:13:08.815786-03:00 polar kernel: [ 2.985024] cdrom: Uniform CD-ROM driver Revision: 3.20 2018-08-28T10:13:08.815790-03:00 polar kernel: [ 2.985298] sr 1:0:0:0: Attached scsi CD-ROM sr0 2018-08-28T10:13:08.815792-03:00 polar kernel: [ 3.058447] usb 1-6: New USB device found, idVendor=413c, idProduct=2113 -- 2018-08-29T17:23:42.925968-03:00 polar kernel: [ 2.939978] ata2.00: ATAPI: PLDS DVD+/-RW DU-8A5LH, 6D1M, max UDMA/133 2018-08-29T17:23:42.925969-03:00 polar kernel: [ 2.944873] ata3: SATA link down (SStatus 4 SControl 300) 2018-08-29T17:23:42.925970-03:00 polar kernel: [ 2.946592] ata2.00: configured for UDMA/133 2018-08-29T17:23:42.925970-03:00 polar kernel: [ 2.963630] scsi 1:0:0:0: CD-ROM PLDS DVD+-RW DU-8A5LH 6D1M PQ: 0 ANSI: 5 2018-08-29T17:23:42.925971-03:00 polar kernel: [ 2.983665] scsi 1:0:0:0: Attached scsi generic sg1 type 5 2018-08-29T17:23:42.925972-03:00 polar kernel: [ 3.007867] sda: sda1 sda2 sda3 sda4 2018-08-29T17:23:42.925972-03:00 polar kernel: [ 3.008600] sd 0:0:0:0: [sda] Attached SCSI disk 2018-08-29T17:23:42.925974-03:00 polar kernel: [ 3.042763] sr 1:0:0:0: [sr0] scsi3-mmc drive: 24x/24x writer cd/rw xa/form2 cdda tray 2018-08-29T17:23:42.925975-03:00 polar kernel: [ 3.042766] cdrom: Uniform CD-ROM driver Revision: 3.20 2018-08-29T17:23:42.925976-03:00 polar kernel: [ 3.043063] sr 1:0:0:0: Attached scsi CD-ROM sr0 2018-08-29T17:23:42.925976-03:00 polar kernel: [ 3.078474] usb 1-6: New USB device found, idVendor=413c, idProduct=2113 -- 2018-08-29T17:29:04.872591-03:00 polar kernel: [ 2.895214] ata2.00: ATAPI: PLDS DVD+/-RW DU-8A5LH, 6D1M, max UDMA/133 2018-08-29T17:29:04.872594-03:00 polar kernel: [ 2.896046] ata2.00: configured for UDMA/133 2018-08-29T17:29:04.872598-03:00 polar kernel: [ 2.903711] scsi 1:0:0:0: CD-ROM PLDS DVD+-RW DU-8A5LH 6D1M PQ: 0 ANSI: 5 2018-08-29T17:29:04.872602-03:00 polar kernel: [ 2.919467] scsi 1:0:0:0: Attached scsi generic sg1 type 5 2018-08-29T17:29:04.872605-03:00 polar kernel: [ 2.936023] usb 1-6: new low-speed USB device number 2 using xhci_hcd 2018-08-29T17:29:04.872608-03:00 polar kernel: [ 2.967264] sda: sda1 sda2 sda3 sda4 2018-08-29T17:29:04.872612-03:00 polar kernel: [ 2.967859] sd 0:0:0:0: [sda] Attached SCSI disk 2018-08-29T17:29:04.872624-03:00 polar kernel: [ 2.998115] sr 1:0:0:0: [sr0] scsi3-mmc drive: 24x/24x writer cd/rw xa/form2 cdda tray 2018-08-29T17:29:04.872626-03:00 polar kernel: [ 2.998118] cdrom: Uniform CD-ROM driver Revision: 3.20 2018-08-29T17:29:04.872628-03:00 polar kernel: [ 2.998343] sr 1:0:0:0: Attached scsi CD-ROM sr0 2018-08-29T17:29:04.872632-03:00 polar kernel: [ 3.094288] usb 1-6: New USB device found, idVendor=413c, idProduct=2113 -- 2018-08-29T17:42:02.904954-03:00 polar kernel: [ 3.053500] ata2: SATA link up 1.5 Gbps (SStatus 113 SControl 300) 2018-08-29T17:42:02.904964-03:00 polar kernel: [ 3.057191] ata2.00: ATAPI: PLDS DVD+/-RW DU-8A5LH, 6D1M, max UDMA/133 2018-08-29T17:42:02.904967-03:00 polar kernel: [ 3.058447] ata2.00: configured for UDMA/133 2018-08-29T17:42:02.904970-03:00 polar kernel: [ 3.070756] scsi 1:0:0:0: CD-ROM PLDS DVD+-RW DU-8A5LH 6D1M PQ: 0 ANSI: 5 2018-08-29T17:42:02.904973-03:00 polar kernel: [ 3.087677] scsi 1:0:0:0: Attached scsi generic sg1 type 5 2018-08-29T17:42:02.904976-03:00 polar kernel: [ 3.101091] sda: sda1 sda2 sda3 sda4 2018-08-29T17:42:02.904979-03:00 polar kernel: [ 3.101807] sd 0:0:0:0: [sda] Attached SCSI disk 2018-08-29T17:42:02.904982-03:00 polar kernel: [ 3.132688] sr 1:0:0:0: [sr0] scsi3-mmc drive: 24x/24x writer cd/rw xa/form2 cdda tray 2018-08-29T17:42:02.904991-03:00 polar kernel: [ 3.132692] cdrom: Uniform CD-ROM driver Revision: 3.20 2018-08-29T17:42:02.904995-03:00 polar kernel: [ 3.132944] sr 1:0:0:0: Attached scsi CD-ROM sr0 2018-08-29T17:42:02.904998-03:00 polar kernel: [ 3.213040] usb 1-6: New USB device found, idVendor=413c, idProduct=2113 -- 2018-08-30T10:21:18.787466-03:00 polar kernel: [ 3.035391] ata2.00: ATAPI: PLDS DVD+/-RW DU-8A5LH, 6D1M, max UDMA/133 2018-08-30T10:21:18.787469-03:00 polar kernel: [ 3.036222] ata2.00: configured for UDMA/133 2018-08-30T10:21:18.787469-03:00 polar kernel: [ 3.043916] scsi 1:0:0:0: CD-ROM PLDS DVD+-RW DU-8A5LH 6D1M PQ: 0 ANSI: 5 2018-08-30T10:21:18.787470-03:00 polar kernel: [ 3.052087] usb 1-6: new low-speed USB device number 2 using xhci_hcd 2018-08-30T10:21:18.787471-03:00 polar kernel: [ 3.063179] scsi 1:0:0:0: Attached scsi generic sg1 type 5 2018-08-30T10:21:18.787472-03:00 polar kernel: [ 3.083566] sda: sda1 sda2 sda3 sda4 2018-08-30T10:21:18.787472-03:00 polar kernel: [ 3.084238] sd 0:0:0:0: [sda] Attached SCSI disk 2018-08-30T10:21:18.787473-03:00 polar kernel: [ 3.113065] sr 1:0:0:0: [sr0] scsi3-mmc drive: 24x/24x writer cd/rw xa/form2 cdda tray 2018-08-30T10:21:18.787475-03:00 polar kernel: [ 3.113068] cdrom: Uniform CD-ROM driver Revision: 3.20 2018-08-30T10:21:18.787476-03:00 polar kernel: [ 3.113272] sr 1:0:0:0: Attached scsi CD-ROM sr0 2018-08-30T10:21:18.787477-03:00 polar kernel: [ 3.213133] usb 1-6: New USB device found, idVendor=413c, idProduct=2113 ########################### I'm trying to run deploy again here, however I'm having some connection issues today (possibly due to the heavy rain) affecting the initial stages of it. If it succeeds, I send the outputs from /var/log/messages on the minions right away. Thanks a lot, Jones On Fri, Aug 31, 2018 at 4:00 AM Eugen Block <ebl...@nde.ag> wrote: > Hi, > > I'm not sure if there's a misunderstanding. You need to track the logs > during the osd deployment step (stage.3), that is where it fails, and > this is where /var/log/messages could be useful. Since the deployment > failed you have no systemd-units (ceph-osd@<ID>.service) to log > anything. > > Before running stage.3 again try something like > > grep -C5 ceph-disk /var/log/messages (or messages-201808*.xz) > > or > > grep -C5 sda4 /var/log/messages (or messages-201808*.xz) > > If that doesn't reveal anything run stage.3 again and watch the logs. > > Regards, > Eugen > > > Zitat von Jones de Andrade <johanne...@gmail.com>: > > > Hi Eugen. > > > > Ok, edited the file /etc/salt/minion, uncommented the "log_level_logfile" > > line and set it to "debug" level. > > > > Turned off the computer, waited a few minutes so that the time frame > would > > stand out in the /var/log/messages file, and restarted the computer. > > > > Using vi I "greped out" (awful wording) the reboot section. From that, I > > also removed most of what it seemed totally unrelated to ceph, salt, > > minions, grafana, prometheus, whatever. > > > > I got the lines below. It does not seem to complain about anything that I > > can see. :( > > > > ################ > > 2018-08-30T15:41:46.455383-03:00 torcello systemd[1]: systemd 234 running > > in system mode. (+PAM -AUDIT +SELINUX -IMA +APPARMOR -SMACK +SYSVINIT > +UTMP > > +LIBCRYPTSETUP +GCRYPT -GNUTLS +ACL +XZ +LZ4 +SECCOMP +BLKID -ELFUTILS > > +KMOD -IDN2 -IDN default-hierarchy=hybrid) > > 2018-08-30T15:41:46.456330-03:00 torcello systemd[1]: Detected > architecture > > x86-64. > > 2018-08-30T15:41:46.456350-03:00 torcello systemd[1]: nss-lookup.target: > > Dependency Before=nss-lookup.target dropped > > 2018-08-30T15:41:46.456357-03:00 torcello systemd[1]: Started Load Kernel > > Modules. > > 2018-08-30T15:41:46.456369-03:00 torcello systemd[1]: Starting Apply > Kernel > > Variables... > > 2018-08-30T15:41:46.457230-03:00 torcello systemd[1]: Started > Alertmanager > > for prometheus. > > 2018-08-30T15:41:46.457237-03:00 torcello systemd[1]: Started Monitoring > > system and time series database. > > 2018-08-30T15:41:46.457403-03:00 torcello systemd[1]: Starting NTP > > client/server... > > > > > > > > > > > > > > *2018-08-30T15:41:46.457425-03:00 torcello systemd[1]: Started Prometheus > > exporter for machine metrics.2018-08-30T15:41:46.457706-03:00 torcello > > prometheus[695]: level=info ts=2018-08-30T18:41:44.797896888Z > > caller=main.go:225 msg="Starting Prometheus" version="(version=2.1.0, > > branch=non-git, revision=non-git)"2018-08-30T15:41:46.457712-03:00 > torcello > > prometheus[695]: level=info ts=2018-08-30T18:41:44.797969232Z > > caller=main.go:226 build_context="(go=go1.9.4, user=abuild@lamb69, > > date=20180513-03:46:03)"2018-08-30T15:41:46.457719-03:00 torcello > > prometheus[695]: level=info ts=2018-08-30T18:41:44.798008802Z > > caller=main.go:227 host_details="(Linux 4.12.14-lp150.12.4-default #1 SMP > > Tue May 22 05:17:22 UTC 2018 (66b2eda) x86_64 torcello > > (none))"2018-08-30T15:41:46.457726-03:00 torcello prometheus[695]: > > level=info ts=2018-08-30T18:41:44.798044088Z caller=main.go:228 > > fd_limits="(soft=1024, hard=4096)"2018-08-30T15:41:46.457738-03:00 > torcello > > prometheus[695]: level=info ts=2018-08-30T18:41:44.802067189Z > > caller=web.go:383 component=web msg="Start listening for connections" > > address=0.0.0.0:9090 <http://0.0.0.0:9090 > >2018-08-30T15:41:46.457745-03:00 > > torcello prometheus[695]: level=info ts=2018-08-30T18:41:44.802037354Z > > caller=main.go:499 msg="Starting TSDB ..."* > > 2018-08-30T15:41:46.458145-03:00 torcello smartd[809]: Monitoring 1 > > ATA/SATA, 0 SCSI/SAS and 0 NVMe devices > > 2018-08-30T15:41:46.458321-03:00 torcello systemd[1]: Started NTP > > client/server. > > *2018-08-30T15:41:50.387157-03:00 torcello ceph_exporter[690]: 2018/08/30 > > 15:41:50 Starting ceph exporter on ":9128"* > > 2018-08-30T15:41:52.658272-03:00 torcello wicked[905]: lo up > > 2018-08-30T15:41:52.658738-03:00 torcello wicked[905]: eth0 up > > 2018-08-30T15:41:52.659989-03:00 torcello systemd[1]: Started wicked > > managed network interfaces. > > 2018-08-30T15:41:52.660514-03:00 torcello systemd[1]: Reached target > > Network. > > 2018-08-30T15:41:52.667938-03:00 torcello systemd[1]: Starting OpenSSH > > Daemon... > > 2018-08-30T15:41:52.668292-03:00 torcello systemd[1]: Reached target > > Network is Online. > > > > > > > > > > *2018-08-30T15:41:52.669132-03:00 torcello systemd[1]: Started Ceph > cluster > > monitor daemon.2018-08-30T15:41:52.669328-03:00 torcello systemd[1]: > > Reached target ceph target allowing to start/stop all ceph-mon@.service > > instances at once.2018-08-30T15:41:52.670346-03:00 torcello systemd[1]: > > Started Ceph cluster manager daemon.2018-08-30T15:41:52.670565-03:00 > > torcello systemd[1]: Reached target ceph target allowing to start/stop > all > > ceph-mgr@.service instances at once.2018-08-30T15:41:52.670839-03:00 > > torcello systemd[1]: Reached target ceph target allowing to start/stop > all > > ceph*@.service instances at once.* > > 2018-08-30T15:41:52.671246-03:00 torcello systemd[1]: Starting Login and > > scanning of iSCSI devices... > > *2018-08-30T15:41:52.672402-03:00 torcello systemd[1]: Starting Grafana > > instance...* > > 2018-08-30T15:41:52.678922-03:00 torcello systemd[1]: Started Backup of > > /etc/sysconfig. > > 2018-08-30T15:41:52.679109-03:00 torcello systemd[1]: Reached target > Timers. > > *2018-08-30T15:41:52.679630-03:00 torcello systemd[1]: Started The Salt > > API.* > > 2018-08-30T15:41:52.692944-03:00 torcello systemd[1]: Starting Postfix > Mail > > Transport Agent... > > *2018-08-30T15:41:52.694687-03:00 torcello systemd[1]: Started The Salt > > Master Server.* > > *2018-08-30T15:41:52.696821-03:00 torcello systemd[1]: Starting The Salt > > Minion...* > > 2018-08-30T15:41:52.772750-03:00 torcello sshd-gen-keys-start[1408]: > > Checking for missing server keys in /etc/ssh > > 2018-08-30T15:41:52.818695-03:00 torcello iscsiadm[1412]: iscsiadm: No > > records found > > 2018-08-30T15:41:52.819541-03:00 torcello systemd[1]: Started Login and > > scanning of iSCSI devices. > > 2018-08-30T15:41:52.820214-03:00 torcello systemd[1]: Reached target > Remote > > File Systems. > > 2018-08-30T15:41:52.821418-03:00 torcello systemd[1]: Starting Permit > User > > Sessions... > > 2018-08-30T15:41:53.045278-03:00 torcello systemd[1]: Started Permit User > > Sessions. > > 2018-08-30T15:41:53.048482-03:00 torcello systemd[1]: Starting Hold until > > boot process finishes up... > > 2018-08-30T15:41:53.054461-03:00 torcello echo[1415]: Starting mail > service > > (Postfix) > > 2018-08-30T15:41:53.447390-03:00 torcello sshd[1431]: Server listening on > > 0.0.0.0 port 22. > > 2018-08-30T15:41:53.447685-03:00 torcello sshd[1431]: Server listening on > > :: port 22. > > 2018-08-30T15:41:53.447907-03:00 torcello systemd[1]: Started OpenSSH > > Daemon. > > > > > > > > > > > > > > > > > > > > > > > > > > > > *2018-08-30T15:41:54.519192-03:00 torcello grafana-server[1413]: > > t=2018-08-30T15:41:54-0300 lvl=info msg="Starting Grafana" logger=server > > version=5.1.3 commit=NA > > compiled=2018-08-30T15:41:53-03002018-08-30T15:41:54.519664-03:00 > torcello > > grafana-server[1413]: t=2018-08-30T15:41:54-0300 lvl=info msg="Config > > loaded from" logger=settings > > file=/usr/share/grafana/conf/defaults.ini2018-08-30T15:41:54.519979-03:00 > > torcello grafana-server[1413]: t=2018-08-30T15:41:54-0300 lvl=info > > msg="Config loaded from" logger=settings > > file=/etc/grafana/grafana.ini2018-08-30T15:41:54.520257-03:00 torcello > > grafana-server[1413]: t=2018-08-30T15:41:54-0300 lvl=info msg="Config > > overridden from command line" logger=settings > > arg="default.paths.data=/var/lib/grafana"2018-08-30T15:41:54.520546-03:00 > > torcello grafana-server[1413]: t=2018-08-30T15:41:54-0300 lvl=info > > msg="Config overridden from command line" logger=settings > > arg="default.paths.logs=/var/log/grafana"2018-08-30T15:41:54.520823-03:00 > > torcello grafana-server[1413]: t=2018-08-30T15:41:54-0300 lvl=info > > msg="Config overridden from command line" logger=settings > > > arg="default.paths.plugins=/var/lib/grafana/plugins"2018-08-30T15:41:54.521085-03:00 > > torcello grafana-server[1413]: t=2018-08-30T15:41:54-0300 lvl=info > > msg="Config overridden from command line" logger=settings > > > arg="default.paths.provisioning=/etc/grafana/provisioning"2018-08-30T15:41:54.521343-03:00 > > torcello grafana-server[1413]: t=2018-08-30T15:41:54-0300 lvl=info > > msg="Path Home" logger=settings > > path=/usr/share/grafana2018-08-30T15:41:54.521593-03:00 torcello > > grafana-server[1413]: t=2018-08-30T15:41:54-0300 lvl=info msg="Path Data" > > logger=settings path=/var/lib/grafana2018-08-30T15:41:54.521843-03:00 > > torcello grafana-server[1413]: t=2018-08-30T15:41:54-0300 lvl=info > > msg="Path Logs" logger=settings > > path=/var/log/grafana2018-08-30T15:41:54.522108-03:00 torcello > > grafana-server[1413]: t=2018-08-30T15:41:54-0300 lvl=info msg="Path > > Plugins" logger=settings > > path=/var/lib/grafana/plugins2018-08-30T15:41:54.522361-03:00 torcello > > grafana-server[1413]: t=2018-08-30T15:41:54-0300 lvl=info msg="Path > > Provisioning" logger=settings > > path=/etc/grafana/provisioning2018-08-30T15:41:54.522611-03:00 torcello > > grafana-server[1413]: t=2018-08-30T15:41:54-0300 lvl=info msg="App mode > > production" logger=settings2018-08-30T15:41:54.522885-03:00 torcello > > grafana-server[1413]: t=2018-08-30T15:41:54-0300 lvl=info msg="Writing > PID > > file" logger=server path=/var/run/grafana/grafana-server.pid pid=1413* > > > > > > > > > > > > > > *2018-08-30T15:41:54.523148-03:00 torcello grafana-server[1413]: > > t=2018-08-30T15:41:54-0300 lvl=info msg="Initializing DB" logger=sqlstore > > dbtype=sqlite32018-08-30T15:41:54.523398-03:00 torcello > > grafana-server[1413]: t=2018-08-30T15:41:54-0300 lvl=info msg="Starting > DB > > migration" logger=migrator2018-08-30T15:41:54.804052-03:00 torcello > > grafana-server[1413]: t=2018-08-30T15:41:54-0300 lvl=info msg="Executing > > migration" logger=migrator id="copy data account to > > org"2018-08-30T15:41:54.804423-03:00 torcello grafana-server[1413]: > > t=2018-08-30T15:41:54-0300 lvl=info msg="Skipping migration condition not > > fulfilled" logger=migrator id="copy data account to > > org"2018-08-30T15:41:54.804724-03:00 torcello grafana-server[1413]: > > t=2018-08-30T15:41:54-0300 lvl=info msg="Executing migration" > > logger=migrator id="copy data account_user to > > org_user"2018-08-30T15:41:54.804985-03:00 torcello grafana-server[1413]: > > t=2018-08-30T15:41:54-0300 lvl=info msg="Skipping migration condition not > > fulfilled" logger=migrator id="copy data account_user to > > org_user"2018-08-30T15:41:54.838327-03:00 torcello grafana-server[1413]: > > t=2018-08-30T15:41:54-0300 lvl=info msg="Starting plugin search" > > logger=plugins* > > 2018-08-30T15:41:54.947408-03:00 torcello systemd[1]: Starting Locale > > Service... > > 2018-08-30T15:41:54.979069-03:00 torcello systemd[1]: Started Locale > > Service. > > > > > > > > > > > > *2018-08-30T15:41:55.023859-03:00 torcello grafana-server[1413]: > > t=2018-08-30T15:41:55-0300 lvl=info msg="Registering plugin" > logger=plugins > > name=Discrete2018-08-30T15:41:55.028462-03:00 torcello > > grafana-server[1413]: t=2018-08-30T15:41:55-0300 lvl=info > msg="Registering > > plugin" logger=plugins name=Monasca2018-08-30T15:41:55.065227-03:00 > > torcello grafana-server[1413]: t=2018-08-30T15:41:55-0300 lvl=eror > > msg="can't read datasource provisioning files from directory" > > logger=provisioning.datasources > > > path=/etc/grafana/provisioning/datasources2018-08-30T15:41:55.065462-03:00 > > torcello grafana-server[1413]: t=2018-08-30T15:41:55-0300 lvl=eror > > msg="can't read dashboard provisioning files from directory" > > logger=provisioning.dashboard > > path=/etc/grafana/provisioning/dashboards2018-08-30T15:41:55.065636-03:00 > > torcello grafana-server[1413]: t=2018-08-30T15:41:55-0300 lvl=info > > msg="Initializing Alerting" > > logger=alerting.engine2018-08-30T15:41:55.065779-03:00 torcello > > grafana-server[1413]: t=2018-08-30T15:41:55-0300 lvl=info > msg="Initializing > > CleanUpService" logger=cleanup* > > 2018-08-30T15:41:55.274779-03:00 torcello systemd[1]: Started Grafana > > instance. > > 2 > > *018-08-30T15:41:55.313056-03:00 torcello grafana-server[1413]: > > t=2018-08-30T15:41:55-0300 lvl=info msg="Initializing Stream > > Manager"2018-08-30T15:41:55.313251-03:00 torcello grafana-server[1413]: > > t=2018-08-30T15:41:55-0300 lvl=info msg="Initializing HTTP Server" > > logger=http.server address=0.0.0.0:3000 <http://0.0.0.0:3000> > protocol=http > > subUrl= socket=* > > 2018-08-30T15:41:58.304749-03:00 torcello systemd[1]: Started Command > > Scheduler. > > 2018-08-30T15:41:58.381694-03:00 torcello systemd[1]: Started The Salt > > Minion. > > 2018-08-30T15:41:58.386643-03:00 torcello cron[1611]: (CRON) INFO > > (RANDOM_DELAY will be scaled with factor 11% if used.) > > 2018-08-30T15:41:58.396087-03:00 torcello cron[1611]: (CRON) INFO > (running > > with inotify support) > > 2018-08-30T15:42:06.367096-03:00 torcello systemd[1]: Started Hold until > > boot process finishes up. > > 2018-08-30T15:42:06.369301-03:00 torcello systemd[1]: Started Getty on > tty1. > > 2018-08-30T15:42:11.535310-03:00 torcello systemd[1792]: Reached target > > Paths. > > 2018-08-30T15:42:11.536128-03:00 torcello systemd[1792]: Starting D-Bus > > User Message Bus Socket. > > 2018-08-30T15:42:11.536378-03:00 torcello systemd[1792]: Reached target > > Timers. > > 2018-08-30T15:42:11.598968-03:00 torcello systemd[1792]: Listening on > D-Bus > > User Message Bus Socket. > > 2018-08-30T15:42:11.599151-03:00 torcello systemd[1792]: Reached target > > Sockets. > > 2018-08-30T15:42:11.599277-03:00 torcello systemd[1792]: Reached target > > Basic System. > > 2018-08-30T15:42:11.599398-03:00 torcello systemd[1792]: Reached target > > Default. > > 2018-08-30T15:42:11.599514-03:00 torcello systemd[1792]: Startup finished > > in 145ms. > > 2018-08-30T15:42:11.599636-03:00 torcello systemd[1]: Started User > Manager > > for UID 464. > > 2018-08-30T15:42:12.471869-03:00 torcello systemd[1792]: Started D-Bus > User > > Message Bus. > > 2018-08-30T15:42:15.898853-03:00 torcello systemd[1]: Starting Disk > > Manager... > > 2018-08-30T15:42:15.974641-03:00 torcello systemd[1]: Started Disk > Manager. > > 2018-08-30T15:42:16.897412-03:00 torcello node_exporter[807]: > > time="2018-08-30T15:42:16-03:00" level=error msg="ERROR: ntp collector > > failed after 0.000087s: couldn't get SNTP reply: read udp 127.0.0.1:42089 > -> > > 127.0.0.1:123: read: connection refused" source="collector.go:123" > > 2018-08-30T15:42:17.589461-03:00 torcello chronyd[845]: Selected source > > 200.189.40.8 > > 2018-08-30T15:43:16.899040-03:00 torcello node_exporter[807]: > > time="2018-08-30T15:43:16-03:00" level=error msg="ERROR: ntp collector > > failed after 0.000105s: couldn't get SNTP reply: read udp 127.0.0.1:59525 > -> > > 127.0.0.1:123: read: connection refused" source="collector.go:123" > > 2018-08-30T15:44:15.496595-03:00 torcello systemd[1792]: Stopped target > > Default. > > 2018-08-30T15:44:15.496824-03:00 torcello systemd[1792]: Stopping D-Bus > > User Message Bus... > > 2018-08-30T15:44:15.502438-03:00 torcello systemd[1792]: Stopped D-Bus > User > > Message Bus. > > 2018-08-30T15:44:15.502627-03:00 torcello systemd[1792]: Stopped target > > Basic System. > > 2018-08-30T15:44:15.502776-03:00 torcello systemd[1792]: Stopped target > > Paths. > > 2018-08-30T15:44:15.502923-03:00 torcello systemd[1792]: Stopped target > > Timers. > > 2018-08-30T15:44:15.503062-03:00 torcello systemd[1792]: Stopped target > > Sockets. > > 2018-08-30T15:44:15.503200-03:00 torcello systemd[1792]: Closed D-Bus > User > > Message Bus Socket. > > 2018-08-30T15:44:15.503356-03:00 torcello systemd[1792]: Reached target > > Shutdown. > > 2018-08-30T15:44:15.503572-03:00 torcello systemd[1792]: Starting Exit > the > > Session... > > 2018-08-30T15:44:15.511298-03:00 torcello systemd[2295]: Starting D-Bus > > User Message Bus Socket. > > 2018-08-30T15:44:15.511493-03:00 torcello systemd[2295]: Reached target > > Timers. > > 2018-08-30T15:44:15.511664-03:00 torcello systemd[2295]: Reached target > > Paths. > > 2018-08-30T15:44:15.517873-03:00 torcello systemd[2295]: Listening on > D-Bus > > User Message Bus Socket. > > 2018-08-30T15:44:15.518060-03:00 torcello systemd[2295]: Reached target > > Sockets. > > 2018-08-30T15:44:15.518216-03:00 torcello systemd[2295]: Reached target > > Basic System. > > 2018-08-30T15:44:15.518373-03:00 torcello systemd[2295]: Reached target > > Default. > > 2018-08-30T15:44:15.518501-03:00 torcello systemd[2295]: Startup finished > > in 31ms. > > 2018-08-30T15:44:15.518634-03:00 torcello systemd[1]: Started User > Manager > > for UID 1000. > > 2018-08-30T15:44:15.518759-03:00 torcello systemd[1792]: Received > > SIGRTMIN+24 from PID 2300 (kill). > > 2018-08-30T15:44:15.537634-03:00 torcello systemd[1]: Stopped User > Manager > > for UID 464. > > 2018-08-30T15:44:15.538422-03:00 torcello systemd[1]: Removed slice User > > Slice of sddm. > > 2018-08-30T15:44:15.613246-03:00 torcello systemd[2295]: Started D-Bus > User > > Message Bus. > > 2018-08-30T15:44:15.623989-03:00 torcello dbus-daemon[2311]: [session > > uid=1000 pid=2311] Successfully activated service > 'org.freedesktop.systemd1' > > 2018-08-30T15:44:16.447162-03:00 torcello kapplymousetheme[2350]: > > kcm_input: Using X11 backend > > 2018-08-30T15:44:16.901642-03:00 torcello node_exporter[807]: > > time="2018-08-30T15:44:16-03:00" level=error msg="ERROR: ntp collector > > failed after 0.000205s: couldn't get SNTP reply: read udp 127.0.0.1:53434 > -> > > 127.0.0.1:123: read: connection refused" source="collector.go:123" > > ################ > > > > Any ideas? > > > > Thanks a lot, > > > > Jones > > > > On Thu, Aug 30, 2018 at 4:14 AM Eugen Block <ebl...@nde.ag> wrote: > > > >> Hi, > >> > >> > So, it only contains logs concerning the node itself (is it correct? > >> sincer > >> > node01 is also the master, I was expecting it to have logs from the > other > >> > too) and, moreover, no ceph-osd* files. Also, I'm looking the logs I > have > >> > available, and nothing "shines out" (sorry for my poor english) as a > >> > possible error. > >> > >> the logging is not configured to be centralised per default, you would > >> have to configure that yourself. > >> > >> Regarding the OSDs, if there are OSD logs created, they're created on > >> the OSD nodes, not on the master. But since the OSD deployment fails, > >> there probably are no OSD specific logs yet. So you'll have to take a > >> look into the syslog (/var/log/messages), that's where the salt-minion > >> reports its attempts to create the OSDs. Chances are high that you'll > >> find the root cause in here. > >> > >> If the output is not enough, set the log-level to debug: > >> > >> osd-1:~ # grep -E "^log_level" /etc/salt/minion > >> log_level: debug > >> > >> > >> Regards, > >> Eugen > >> > >> > >> Zitat von Jones de Andrade <johanne...@gmail.com>: > >> > >> > Hi Eugen. > >> > > >> > Sorry for the delay in answering. > >> > > >> > Just looked in the /var/log/ceph/ directory. It only contains the > >> following > >> > files (for example on node01): > >> > > >> > ####### > >> > # ls -lart > >> > total 3864 > >> > -rw------- 1 ceph ceph 904 ago 24 13:11 ceph.audit.log-20180829.xz > >> > drwxr-xr-x 1 root root 898 ago 28 10:07 .. > >> > -rw-r--r-- 1 ceph ceph 189464 ago 28 23:59 > >> ceph-mon.node01.log-20180829.xz > >> > -rw------- 1 ceph ceph 24360 ago 28 23:59 ceph.log-20180829.xz > >> > -rw-r--r-- 1 ceph ceph 48584 ago 29 00:00 > >> ceph-mgr.node01.log-20180829.xz > >> > -rw------- 1 ceph ceph 0 ago 29 00:00 ceph.audit.log > >> > drwxrws--T 1 ceph ceph 352 ago 29 00:00 . > >> > -rw-r--r-- 1 ceph ceph 1908122 ago 29 12:46 ceph-mon.node01.log > >> > -rw------- 1 ceph ceph 175229 ago 29 12:48 ceph.log > >> > -rw-r--r-- 1 ceph ceph 1599920 ago 29 12:49 ceph-mgr.node01.log > >> > ####### > >> > > >> > So, it only contains logs concerning the node itself (is it correct? > >> sincer > >> > node01 is also the master, I was expecting it to have logs from the > other > >> > too) and, moreover, no ceph-osd* files. Also, I'm looking the logs I > have > >> > available, and nothing "shines out" (sorry for my poor english) as a > >> > possible error. > >> > > >> > Any suggestion on how to proceed? > >> > > >> > Thanks a lot in advance, > >> > > >> > Jones > >> > > >> > > >> > On Mon, Aug 27, 2018 at 5:29 AM Eugen Block <ebl...@nde.ag> wrote: > >> > > >> >> Hi Jones, > >> >> > >> >> all ceph logs are in the directory /var/log/ceph/, each daemon has > its > >> >> own log file, e.g. OSD logs are named ceph-osd.*. > >> >> > >> >> I haven't tried it but I don't think SUSE Enterprise Storage deploys > >> >> OSDs on partitioned disks. Is there a way to attach a second disk to > >> >> the OSD nodes, maybe via USB or something? > >> >> > >> >> Although this thread is ceph related it is referring to a specific > >> >> product, so I would recommend to post your question in the SUSE forum > >> >> [1]. > >> >> > >> >> Regards, > >> >> Eugen > >> >> > >> >> [1] > https://forums.suse.com/forumdisplay.php?99-SUSE-Enterprise-Storage > >> >> > >> >> Zitat von Jones de Andrade <johanne...@gmail.com>: > >> >> > >> >> > Hi Eugen. > >> >> > > >> >> > Thanks for the suggestion. I'll look for the logs (since it's our > >> first > >> >> > attempt with ceph, I'll have to discover where they are, but no > >> problem). > >> >> > > >> >> > One thing called my attention on your response however: > >> >> > > >> >> > I haven't made myself clear, but one of the failures we encountered > >> were > >> >> > that the files now containing: > >> >> > > >> >> > node02: > >> >> > ---------- > >> >> > storage: > >> >> > ---------- > >> >> > osds: > >> >> > ---------- > >> >> > /dev/sda4: > >> >> > ---------- > >> >> > format: > >> >> > bluestore > >> >> > standalone: > >> >> > True > >> >> > > >> >> > Were originally empty, and we filled them by hand following a model > >> found > >> >> > elsewhere on the web. It was necessary, so that we could continue, > but > >> >> the > >> >> > model indicated that, for example, it should have the path for > >> /dev/sda > >> >> > here, not /dev/sda4. We chosen to include the specific partition > >> >> > identification because we won't have dedicated disks here, rather > just > >> >> the > >> >> > very same partition as all disks were partitioned exactly the same. > >> >> > > >> >> > While that was enough for the procedure to continue at that point, > >> now I > >> >> > wonder if it was the right call and, if it indeed was, if it was > done > >> >> > properly. As such, I wonder: what you mean by "wipe" the partition > >> here? > >> >> > /dev/sda4 is created, but is both empty and unmounted: Should a > >> different > >> >> > operation be performed on it, should I remove it first, should I > have > >> >> > written the files above with only /dev/sda as target? > >> >> > > >> >> > I know that probably I wouldn't run in this issues with dedicated > >> discks, > >> >> > but unfortunately that is absolutely not an option. > >> >> > > >> >> > Thanks a lot in advance for any comments and/or extra suggestions. > >> >> > > >> >> > Sincerely yours, > >> >> > > >> >> > Jones > >> >> > > >> >> > On Sat, Aug 25, 2018 at 5:46 PM Eugen Block <ebl...@nde.ag> wrote: > >> >> > > >> >> >> Hi, > >> >> >> > >> >> >> take a look into the logs, they should point you in the right > >> direction. > >> >> >> Since the deployment stage fails at the OSD level, start with the > OSD > >> >> >> logs. Something's not right with the disks/partitions, did you > wipe > >> >> >> the partition from previous attempts? > >> >> >> > >> >> >> Regards, > >> >> >> Eugen > >> >> >> > >> >> >> Zitat von Jones de Andrade <johanne...@gmail.com>: > >> >> >> > >> >> >>> (Please forgive my previous email: I was using another message > and > >> >> >>> completely forget to update the subject) > >> >> >>> > >> >> >>> Hi all. > >> >> >>> > >> >> >>> I'm new to ceph, and after having serious problems in ceph stages > >> 0, 1 > >> >> >> and > >> >> >>> 2 that I could solve myself, now it seems that I have hit a wall > >> harder > >> >> >>> than my head. :) > >> >> >>> > >> >> >>> When I run salt-run state.orch ceph.stage.deploy, i monitor I > see it > >> >> >> going > >> >> >>> up to here: > >> >> >>> > >> >> >>> ####### > >> >> >>> [14/71] ceph.sysctl on > >> >> >>> node01....................................... ✓ (0.5s) > >> >> >>> node02........................................ ✓ (0.7s) > >> >> >>> node03....................................... ✓ (0.6s) > >> >> >>> node04......................................... ✓ > (0.5s) > >> >> >>> node05....................................... ✓ (0.6s) > >> >> >>> node06.......................................... ✓ > (0.5s) > >> >> >>> > >> >> >>> [15/71] ceph.osd on > >> >> >>> node01...................................... ❌ (0.7s) > >> >> >>> node02........................................ ❌ (0.7s) > >> >> >>> node03....................................... ❌ (0.7s) > >> >> >>> node04......................................... ❌ > (0.6s) > >> >> >>> node05....................................... ❌ (0.6s) > >> >> >>> node06.......................................... ❌ > (0.7s) > >> >> >>> > >> >> >>> Ended stage: ceph.stage.deploy succeeded=14/71 failed=1/71 > >> time=624.7s > >> >> >>> > >> >> >>> Failures summary: > >> >> >>> > >> >> >>> ceph.osd (/srv/salt/ceph/osd): > >> >> >>> node02: > >> >> >>> deploy OSDs: Module function osd.deploy threw an exception. > >> >> >> Exception: > >> >> >>> Mine on node02 for cephdisks.list > >> >> >>> node03: > >> >> >>> deploy OSDs: Module function osd.deploy threw an exception. > >> >> >> Exception: > >> >> >>> Mine on node03 for cephdisks.list > >> >> >>> node01: > >> >> >>> deploy OSDs: Module function osd.deploy threw an exception. > >> >> >> Exception: > >> >> >>> Mine on node01 for cephdisks.list > >> >> >>> node04: > >> >> >>> deploy OSDs: Module function osd.deploy threw an exception. > >> >> >> Exception: > >> >> >>> Mine on node04 for cephdisks.list > >> >> >>> node05: > >> >> >>> deploy OSDs: Module function osd.deploy threw an exception. > >> >> >> Exception: > >> >> >>> Mine on node05 for cephdisks.list > >> >> >>> node06: > >> >> >>> deploy OSDs: Module function osd.deploy threw an exception. > >> >> >> Exception: > >> >> >>> Mine on node06 for cephdisks.list > >> >> >>> ####### > >> >> >>> > >> >> >>> Since this is a first attempt in 6 simple test machines, we are > >> going > >> >> to > >> >> >>> put the mon, osds, etc, in all nodes at first. Only the master is > >> left > >> >> >> in a > >> >> >>> single machine (node01) by now. > >> >> >>> > >> >> >>> As they are simple machines, they have a single hdd, which is > >> >> partitioned > >> >> >>> as follows (the hda4 partition is unmounted and left for the ceph > >> >> >> system): > >> >> >>> > >> >> >>> ########### > >> >> >>> # lsblk > >> >> >>> NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT > >> >> >>> sda 8:0 0 465,8G 0 disk > >> >> >>> ├─sda1 8:1 0 500M 0 part /boot/efi > >> >> >>> ├─sda2 8:2 0 16G 0 part [SWAP] > >> >> >>> ├─sda3 8:3 0 49,3G 0 part / > >> >> >>> └─sda4 8:4 0 400G 0 part > >> >> >>> sr0 11:0 1 3,7G 0 rom > >> >> >>> > >> >> >>> # salt -I 'roles:storage' cephdisks.list > >> >> >>> node01: > >> >> >>> node02: > >> >> >>> node03: > >> >> >>> node04: > >> >> >>> node05: > >> >> >>> node06: > >> >> >>> > >> >> >>> # salt -I 'roles:storage' pillar.get ceph > >> >> >>> node02: > >> >> >>> ---------- > >> >> >>> storage: > >> >> >>> ---------- > >> >> >>> osds: > >> >> >>> ---------- > >> >> >>> /dev/sda4: > >> >> >>> ---------- > >> >> >>> format: > >> >> >>> bluestore > >> >> >>> standalone: > >> >> >>> True > >> >> >>> (and so on for all 6 machines) > >> >> >>> ########## > >> >> >>> > >> >> >>> Finally and just in case, my policy.cfg file reads: > >> >> >>> > >> >> >>> ######### > >> >> >>> #cluster-unassigned/cluster/*.sls > >> >> >>> cluster-ceph/cluster/*.sls > >> >> >>> profile-default/cluster/*.sls > >> >> >>> profile-default/stack/default/ceph/minions/*yml > >> >> >>> config/stack/default/global.yml > >> >> >>> config/stack/default/ceph/cluster.yml > >> >> >>> role-master/cluster/node01.sls > >> >> >>> role-admin/cluster/*.sls > >> >> >>> role-mon/cluster/*.sls > >> >> >>> role-mgr/cluster/*.sls > >> >> >>> role-mds/cluster/*.sls > >> >> >>> role-ganesha/cluster/*.sls > >> >> >>> role-client-nfs/cluster/*.sls > >> >> >>> role-client-cephfs/cluster/*.sls > >> >> >>> ########## > >> >> >>> > >> >> >>> Please, could someone help me and shed some light on this issue? > >> >> >>> > >> >> >>> Thanks a lot in advance, > >> >> >>> > >> >> >>> Regasrds, > >> >> >>> > >> >> >>> Jones > >> >> >> > >> >> >> > >> >> >> > >> >> >> _______________________________________________ > >> >> >> ceph-users mailing list > >> >> >> ceph-users@lists.ceph.com > >> >> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >> >> >> > >> >> > >> >> > >> >> > >> > >> > >> > >> _______________________________________________ > >> ceph-users mailing list > >> ceph-users@lists.ceph.com > >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >> > > > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com