Hi

Why is "ceph -s" showing only two monitors while three monitor services are 
running ?

# ceph versions
{   "mon": {        "ceph version 16.2.5 
(0883bdea7337b95e4b611c768c0279868462204a) pacific (stable)": 2     },
    "mgr": {         "ceph version 16.2.5 
(0883bdea7337b95e4b611c768c0279868462204a) pacific (stable)": 2     },
    "osd": {         "ceph version 16.2.5 
(0883bdea7337b95e4b611c768c0279868462204a) pacific (stable)": 36     },
    "mds": {         "ceph version 16.2.5 
(0883bdea7337b95e4b611c768c0279868462204a) pacific (stable)": 1     },
    "rgw": {         "ceph version 16.2.5 
(0883bdea7337b95e4b611c768c0279868462204a) pacific (stable)": 3     },
    "overall": {    "ceph version 16.2.5 
(0883bdea7337b95e4b611c768c0279868462204a) pacific (stable)": 44     }     }

# ceph orch ls
NAME             PORTS                  RUNNING  REFRESHED  AGE  PLACEMENT
crash                                       3/3  8m ago     2y   label:ceph
ingress.nfs.nfs  10.45.128.8:2049,9049      4/4  8m ago     2y   count:2
mds.cephfs                                  3/3  8m ago     2y   
count:3;label:mdss
mgr                                         3/3  8m ago     23M  
a001s016;a001s017;a001s018;count:3
mon                                         3/3  8m ago     16h  
a001s016;a001s017;a001s018;count:3   <== [ 3 monitor services running]
nfs.nfs          ?:12049                    3/3  8m ago     2y   
a001s016;a001s017;a001s018;count:3
node-exporter    ?:9100                     3/3  8m ago     2y   *
osd.unmanaged                             36/36  8m ago     -    <unmanaged>
prometheus       ?:9095                     1/1  10s ago    23M  count:1
rgw.ceph         ?:8080                     3/3  8m ago     19h  
count-per-host:1;label:rgws
root@a001s017:~# ceph -s
  cluster:
    id:     604d56db-2fab-45db-a9ea-c418f9a8cca8
    health: HEALTH_OK

  services:
    mon: 2 daemons, quorum a001s018,a001s017 (age 16h)  <== [ shows ONLY 2 
monitors running]
    mgr: a001s017.bpygfm(active, since 13M), standbys: a001s016.ctmoay
    mds: 1/1 daemons up, 2 standby
    osd: 36 osds: 36 up (since 54s), 36 in (since 2y)
    rgw: 3 daemons active (3 hosts, 1 zones)

  data:
    volumes: 1/1 healthy
    pools:   43 pools, 1633 pgs
    objects: 51.81M objects, 77 TiB
    usage:   120 TiB used, 131 TiB / 252 TiB avail
    pgs:     1631 active+clean
             2    active+clean+scrubbing+deep

  io:
    client:   220 MiB/s rd, 448 MiB/s wr, 251 op/s rd, 497 op/s wr

# ceph orch ls --service_name=mon
NAME  PORTS  RUNNING  REFRESHED  AGE  PLACEMENT
mon              3/3  8m ago     16h  a001s016;a001s017;a001s018;count:3  <== [ 
3 monitors running ]

# ceph orch ps --daemon_type=mon
NAME          HOST      PORTS  STATUS         REFRESHED  AGE  MEM USE  MEM LIM  
VERSION  IMAGE ID      CONTAINER ID
mon.a001s016  a001s016         running (19h)     9m ago  19h     706M    2048M  
16.2.5   6e73176320aa  8484a912f96a
mon.a001s017  a001s017         running (16h)    66s ago  19h     949M    2048M  
16.2.5   6e73176320aa  e5e5cb6c256c   <== [ 3  mon daemons running ]
mon.a001s018  a001s018         running (5w)      2m ago   2y    1155M    2048M  
16.2.5   6e73176320aa  7d2bb6d41f54

a001s016# systemctl --type=service | grep @mon
  
ceph-604d56db-2fab-45db-a9ea-c418f9a8cca8@mon.a001s016.service<mailto:ceph-604d56db-2fab-45db-a9ea-c418f9a8cca8@mon.a001s016.service>
                loaded active running Ceph mon.a001s016 for 
604d56db-2fab-45db-a9ea-c418f9a8cca8
a001s017# systemctl --type=service | grep @mon
  
ceph-604d56db-2fab-45db-a9ea-c418f9a8cca8@mon.a001s017.service<mailto:ceph-604d56db-2fab-45db-a9ea-c418f9a8cca8@mon.a001s017.service>
                loaded active running Ceph mon.a001s017 for 
604d56db-2fab-45db-a9ea-c418f9a8cca8
a001s018# systemctl --type=service | grep @mon
  
ceph-604d56db-2fab-45db-a9ea-c418f9a8cca8@mon.a001s018.service<mailto:ceph-604d56db-2fab-45db-a9ea-c418f9a8cca8@mon.a001s018.service>
                loaded active running Ceph mon.a001s018 for 
604d56db-2fab-45db-a9ea-c418f9a8cca8


Thank you,
Anantha
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to