I tried setting noout and that did provide a bit better result.
Basically I could stop the OSD on the inactive server and everything
still worked (after a 2-3 second pause) but then when I rebooted the
inactive server everything hung again until it came back online and
resynced with the cluster. This is what I saw in ceph -s:
cluster eb2003cf-b16d-4551-adb7-892469447f89
health HEALTH_WARN
128 pgs degraded
124 pgs stuck unclean
128 pgs undersized
recovery 805252/1610504 objects degraded (50.000%)
mds cluster is degraded
1/2 in osds are down
noout flag(s) set
monmap e1: 3 mons at
{FILE1=10.1.1.201:6789/0,FILE2=10.1.1.202:6789/0,MON1=10.1.1.90:6789/0}
election epoch 216, quorum 0,1,2 FILE1,FILE2,MON1
fsmap e796: 1/1/1 up {0=FILE2=up:rejoin}
osdmap e360: 2 osds: 1 up, 2 in; 128 remapped pgs
flags noout,sortbitwise,require_jewel_osds
pgmap v7056802: 128 pgs, 3 pools, 164 GB data, 786 kobjects
349 GB used, 550 GB / 899 GB avail
805252/1610504 objects degraded (50.000%)
128 active+undersized+degraded
client io 1379 B/s rd, 1 op/s rd, 0 op/s wr
These are the commands I ran and the results:
ceph osd set noout
systemctl stop ceph-mds@FILE2.service
# Everything still works on the clients...
systemctl stop ceph-osd@0.service # This was on FILE2 wile FILE1 was the
active fsmap
# Fails over quickly, can still read content on the clients..
# Rebooted FILE2
# File access on the clients locked up until FILE2 rejoined
This is on Ubuntu 16 with kernel 4.4.0-141, so I'm not sure if that
qualifies for David's warning about old kernels...
Is there a command or a logfile I can look at that will better help to
diagnose this issue? Is three servers (with only 2 OSDs) enough to run
a HA cluster on ceph, or does it just die when it doesn't have 3 active
servers for a quorum? Would installing MDS and MON on a 4th box (but
sticking with 2 OSDs) be what's required to resolve this? I really
don't want to do that, but if I have to I guess I can look into find
another box.
On 2019-01-21 5:01 p.m., ceph-users-requ...@lists.ceph.com wrote:
Message: 14
Date: Mon, 21 Jan 2019 10:05:15 +0100
From: Robert Sander<r.san...@heinlein-support.de>
To:ceph-users@lists.ceph.com
Subject: Re: [ceph-users] How To Properly Failover a HA Setup
Message-ID:<587dac75-96bc-8719-ee62-38e71491c...@heinlein-support.de>
Content-Type: text/plain; charset="utf-8"
On 21.01.19 09:22, Charles Tassell wrote:
Hello Everyone,
? I've got a 3 node Jewel cluster setup, and I think I'm missing
something.? When I want to take one of my nodes down for maintenance
(kernel upgrades or the like) all of my clients (running the kernel
module for the cephfs filesystem) hang for a couple of minutes before
the redundant servers kick in.
Have you set the noout flag before doing cluster maintenance?
ceph osd set noout
and afterwards
ceph osd unset noout
Regards
-- Robert Sander Heinlein Support GmbH Schwedter Str. 8/9b, 10119 Berlin
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com