[ceph-users] Re: Ceph osd df tree takes a long time to respond

2024-06-20 Thread Huy Nguyen
Thanks, I just found out the issue
https://tracker.ceph.com/issues/43317
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: wrong public_ip after blackout / poweroutage

2024-06-20 Thread Eugen Block

Hi,

this only a theory, not a proven answer or something. But the  
orchestrator does automatically reconfigure daemons depending on the  
circumstances. So my theory is, some of the OSD nodes didn't respond  
via public network anymore, so ceph tried to use the cluster network  
as a fallback. The other way around is more common: if you don't have  
a cluster network configured at all, you see logs stating "falling  
back to public interface" (or similar). If the orchestrator did  
reconfigure the daemons, it would have been logged in the active mgr.  
And the result would be a different ceph.conf for the daemons in  
/var/lib/ceph/{FSID}/osd.{OSD_ID}/config. If you still have the mgr  
logs from after the outage you might find some clues.


Regards,
Eugen

Zitat von mailing-lists :


Dear Cephers,

after a succession of unfortunate events, we have suffered a  
complete datacenter blackout today.



Ceph _nearly_ perfectly came back up. The Health was OK and all  
services were online, but we were having weird problems. Weird as  
in, we could sometimes map rbds and sometimes not, and sometimes we  
could use the cephfs and sometimes we could not...


Turns out, some osds (id say 5%) came back with the cluster_ip  
address as their public_ip and thus were not reachable.


I do not see any pattern, why some osds are faulty and others are  
not. The fault is spread over nearly all nodes. This is an example:


osd.45 up   in  weight 1 up_from 184143 up_thru 184164 down_at  
184142 last_clean_interval [182655,184103)  
[v2:192.168.222.20:6834/1536394698,v1:192.168.222.20:6842/1536394698]  
[v2:192.168.222.20:6848/1536394698,v1:192.168.222.20:6853/1536394698]  
exists,up 002326c9


This should have a public_ip in the first brackets []. Our  
cluster-network is 192.168.222.0/24, which is of course only  
available on the ceph internal switch.


Simply restarting the osds that were affected solved this problem...  
So I am not really asking for your help troubleshooting this; I  
would just like to understand if there is a reasonable explanation.


My guess would be some kind of race-condition when the interfaces  
came up, but then again, why on ~5% of all osds? ... Anyways im  
tired, I hope that this mail is somewhat understandable.



We are running Ceph 17.2.7 with cephadm docker deployment.


If you have any ideas for the cause of this, please let me know. I  
have not seen this issue when I'm gracefully rebooting the nodes.



Best

___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io



___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Full list of metrics provided by ceph exporter daemon

2024-06-20 Thread Anthony D'Atri


> 
> That's not the answer I expected to see :( Are you suggesting enabling all
> possible daemons of Ceph just to obtain a full list of metrics?

Depends if what you're after is potential metrics or ones currently surfaced, I 
assumed you meant the latter.

> Isn't there any list in ceph repo or in documentation with all metrics 
> exposed?

Use the source, Luke?  I don't know of a documented list.

> Another question: ceph_rgw_qactive metric changed its instance_id from
> numeric format to a letter format. Is there any pull request for that in
> Ceph or it is Rook initiative? Example:

I believe recently there was a restructuring of some of the RGW metrics.

> 
> metrics from prometheus module:
> ceph_rgw_qactive{instance_id="4134960"} 0.0
> ceph_rgw_qactive{instance_id="4493203"} 0.0
> 
> metrics from ceph-exporter:
> ceph_rgw_qactive{instance_id="a"} 0 ceph_rgw_qactive{instance_id="a"} 0
> 
> 
> чт, 20 июн. 2024 г. в 20:09, Anthony D'Atri :
> 
>> curl http://endpoint:port/metrics
>> 
>>> On Jun 20, 2024, at 10:15, Peter Razumovsky 
>> wrote:
>>> 
>>> Hello!
>>> 
>>> I'm using Ceph Reef with Rook v1.13 and want to find somewhere a full
>> list
>>> of metrics exported by brand new ceph exporter daemon. We found that some
>>> metrics have been changed after moving from prometheus module metrics to
>> a
>>> separate daemon.
>>> 
>>> We used described method [1] to give us a time to mitigate all risks
>> during
>>> transfer to a new daemon. Now we want to obtain a full list of
>>> ceph-exporter metrics to compare old and new metrics and to mitigate
>>> potential risks of losing monitoring data and breaking alerting systems.
>> We
>>> found some examples of metrics [2] but it is not complete so we will
>>> appreciate it if someone points us to the full list.
>>> 
>>> [1]
>>> 
>> https://docs.ceph.com/en/latest/mgr/prometheus/#ceph-daemon-performance-counters-metrics
>>> 
>>> [2] https://docs.ceph.com/en/latest/monitoring/
>>> 
>>> --
>>> Best regards,
>>> Peter Razumovsky
>>> ___
>>> ceph-users mailing list -- ceph-users@ceph.io
>>> To unsubscribe send an email to ceph-users-le...@ceph.io
>> 
>> 
> 
> -- 
> Best regards,
> Peter Razumovsky
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Full list of metrics provided by ceph exporter daemon

2024-06-20 Thread Peter Razumovsky
That's not the answer I expected to see :( Are you suggesting enabling all
possible daemons of Ceph just to obtain a full list of metrics? Isn't there
any list in ceph repo or in documentation with all metrics exposed?

Another question: ceph_rgw_qactive metric changed its instance_id from
numeric format to a letter format. Is there any pull request for that in
Ceph or it is Rook initiative? Example:

metrics from prometheus module:
ceph_rgw_qactive{instance_id="4134960"} 0.0
ceph_rgw_qactive{instance_id="4493203"} 0.0

metrics from ceph-exporter:
ceph_rgw_qactive{instance_id="a"} 0 ceph_rgw_qactive{instance_id="a"} 0


чт, 20 июн. 2024 г. в 20:09, Anthony D'Atri :

> curl http://endpoint:port/metrics
>
> > On Jun 20, 2024, at 10:15, Peter Razumovsky 
> wrote:
> >
> > Hello!
> >
> > I'm using Ceph Reef with Rook v1.13 and want to find somewhere a full
> list
> > of metrics exported by brand new ceph exporter daemon. We found that some
> > metrics have been changed after moving from prometheus module metrics to
> a
> > separate daemon.
> >
> > We used described method [1] to give us a time to mitigate all risks
> during
> > transfer to a new daemon. Now we want to obtain a full list of
> > ceph-exporter metrics to compare old and new metrics and to mitigate
> > potential risks of losing monitoring data and breaking alerting systems.
> We
> > found some examples of metrics [2] but it is not complete so we will
> > appreciate it if someone points us to the full list.
> >
> > [1]
> >
> https://docs.ceph.com/en/latest/mgr/prometheus/#ceph-daemon-performance-counters-metrics
> >
> > [2] https://docs.ceph.com/en/latest/monitoring/
> >
> > --
> > Best regards,
> > Peter Razumovsky
> > ___
> > ceph-users mailing list -- ceph-users@ceph.io
> > To unsubscribe send an email to ceph-users-le...@ceph.io
>
>

-- 
Best regards,
Peter Razumovsky
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Full list of metrics provided by ceph exporter daemon

2024-06-20 Thread Anthony D'Atri
curl http://endpoint:port/metrics

> On Jun 20, 2024, at 10:15, Peter Razumovsky  wrote:
> 
> Hello!
> 
> I'm using Ceph Reef with Rook v1.13 and want to find somewhere a full list
> of metrics exported by brand new ceph exporter daemon. We found that some
> metrics have been changed after moving from prometheus module metrics to a
> separate daemon.
> 
> We used described method [1] to give us a time to mitigate all risks during
> transfer to a new daemon. Now we want to obtain a full list of
> ceph-exporter metrics to compare old and new metrics and to mitigate
> potential risks of losing monitoring data and breaking alerting systems. We
> found some examples of metrics [2] but it is not complete so we will
> appreciate it if someone points us to the full list.
> 
> [1]
> https://docs.ceph.com/en/latest/mgr/prometheus/#ceph-daemon-performance-counters-metrics
> 
> [2] https://docs.ceph.com/en/latest/monitoring/
> 
> -- 
> Best regards,
> Peter Razumovsky
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Daily slow ops around the same time on different osds

2024-06-20 Thread Eugen Block

Hi,


around the same time


sounds like deep-scrubbing. Did you verify if those OSDs from the  
mentioned pool were scrubbed around that time? The primary OSDs for  
the PGs would log that. Is that pool that heavily used? Or could there  
be one failing disk?


Zitat von "Szabo, Istvan (Agoda)" :


Hi,

For some reason daily I receive a slow ops which affect the rgw.log  
pool the most (which is pretty small, 32pg / 57k objects / 9.5GB data)


Could be the issue related to rocksdb tasks?

Some log lines:
...
2024-06-18T14:17:43.849+0700 7fd8002ea700  1 heartbeat_map  
reset_timeout 'OSD::osd_op_tp thread 0x7fd8002ea700' had timed out  
after 15
2024-06-18T14:17:50.865+0700 7fd81a31e700 -1 osd.461 472771  
get_health_metrics reporting 1 slow ops, oldest is  
osd_op(client.3349483990.0:462716190 22.11  
22:8deed178:::meta.log.08e85e9f-c16e-43f0-b88d-362f3b7ced2d.15:head  
[call log.list in=69b] snapc 0=[] ondisk+read+known_if_redirected  
e472771)
2024-06-18T14:17:50.865+0700 7fd81a31e700  0 log_channel(cluster)  
log [WRN] : 1 slow requests (by type [ 'delayed' : 1 ] most affected  
pool [ 'dc.rgw.log' : 1 ])
2024-06-18T14:19:36.040+0700 7fd81e5b4700  1 heartbeat_map  
is_healthy 'OSD::osd_op_tp

...
2024-06-18T14:19:51.908+0700 7fd81a31e700 -1 osd.461 472771  
get_health_metrics reporting 2 slow ops, oldest is  
osd_op(client.3371539359.0:5095510 22.11  
22:8df1b9e1:::datalog.sync-status.shard.61c9d940-fde4-4bed-9389-edc8d7741817.111:
head [call lock.lock in=64b] snapc 0=[]  
ondisk+write+known_if_redirected e472771)
2024-06-18T14:19:51.908+0700 7fd81a31e700  0 log_channel(cluster)  
log [WRN] : 2 slow requests (by type [ 'delayed' : 2 ] most affected  
pool [ 'dc.rgw.log' : 2 ])
2024-06-18T14:19:52.899+0700 7fd81a31e700 -1 osd.461 472771  
get_health_metrics reporting 2 slow ops, oldest is  
osd_op(client.3371539359.0:5095510 22.11  
22:8df1b9e1:::datalog.sync-status.shard.61c9d940-fde4-4bed-9389-edc8d7741817.111:


I found this thread:  
https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/thread/FR5V466HBXGRVL3Z3RAFKUPQ2FGK2T53/

but I think it is not relevant for me.

We use only SSDs without separated rocksdb so we don' have spillover.
CPU is not overloaded (80% idle) however it's interesting that in  
vmstat the "r" is pretty big which indicates waiting list for cpu:


procs ---memory-- ---swap-- -io -system--  
--cpu-
 r  b   swpd   free   buff  cache   si   sobibo   in   cs us  
sy id wa st
12  0  0 6205288 329803136 1563384400  2149  18940
 0  5  3 92  0  0
17  0  0 6156780 329827840 1563725200 125584 250416  
646098 1965051 10  8 82  0  0
21  0  0 6154636 329849504 1563611200 99320 245024  
493324 1682377  7  6 87  0  0
16  0  0 6144256 329869664 1563692400 87484 301968  
623345 1993374  8  7 84  0  0
19  0  0 6057012 329890080 1563793200 160444 303664  
549194 1820562  8  6 85  0  0


Any idea what this could indicate on octopus 15.2.17 with ubuntu 20.04?

ty



This message is confidential and is for the sole use of the intended  
recipient(s). It may also be privileged or otherwise protected by  
copyright or other legal rules. If you have received it by mistake  
please let us know by reply email and delete it from your system. It  
is prohibited to copy this message or disclose its content to  
anyone. Any confidentiality or privilege is not waived or lost by  
any mistaken delivery or unauthorized disclosure of the message. All  
messages sent to and from Agoda may be monitored to ensure  
compliance with company policies, to protect the company's interests  
and to remove potential malware. Electronic messages may be  
intercepted, amended, lost or deleted, or contain viruses.

___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io



___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Replacing SSD disk(metadata, rocksdb) which are associated with HDDs(osd block)

2024-06-20 Thread Eugen Block

Hi,

- Is it common practice to configure SSDs with block.db and  
associate them with five HDD disks to store OSD blocks when using  
eight SSDs and forty HDDs?


yes, it is common practice.

  Or would it be better to only store the rgw index on SSDs? I am  
also curious about the difference in performance between these  
configurations.


With performance the response is usually "it depends". If performance  
is an issue, why not go all flash?


- If SSDs are configured with block.db as described above, will it  
be necessary to reinstall the five associated OSDs (HDDs) if one SSD  
fails ?


Yes, most likely all OSDs depending on that single SSD will have to be  
redeployed. There's a chance you might be able to rescue some of the  
rocksDB LVs (pvmove)


Regards,
Eugen


Zitat von TaekSoo Lim :

Hi all, I have configured a ceph cluster to be used as an object  
store with a combination of SSDs and HDDs, where the block.db is  
stored on LVM on SSDs and the OSD block is stored on HDDs.
I have set up one SSD for storing metadata (rocksDB), and five HDDs  
are associated with it to store the OSD blocks.
During a disk replacement test, if one SSD that stores the block.db  
fails, all five associated OSDs go down.
When replacing and recovering the failed SSD, it seems necessary to  
reconfigure the five OSDs, which takes too long to recover data and  
has significant performance impact.

So here are two questions:

- Is it common practice to configure SSDs with block.db and  
associate them with five HDD disks to store OSD blocks when using  
eight SSDs and forty HDDs?
  Or would it be better to only store the rgw index on SSDs? I am  
also curious about the difference in performance between these  
configurations.
- If SSDs are configured with block.db as described above, will it  
be necessary to reinstall the five associated OSDs (HDDs) if one SSD  
fails ?
  Also, is there a way to reconstruct the metadata on the newly  
replaced SSD from the remaining five intact HDDs?


As a novice in ceph clusters, I seek advice from experienced users.
Thank you.
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io



___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Full list of metrics provided by ceph exporter daemon

2024-06-20 Thread Peter Razumovsky
Hello!

I'm using Ceph Reef with Rook v1.13 and want to find somewhere a full list
of metrics exported by brand new ceph exporter daemon. We found that some
metrics have been changed after moving from prometheus module metrics to a
separate daemon.

We used described method [1] to give us a time to mitigate all risks during
transfer to a new daemon. Now we want to obtain a full list of
ceph-exporter metrics to compare old and new metrics and to mitigate
potential risks of losing monitoring data and breaking alerting systems. We
found some examples of metrics [2] but it is not complete so we will
appreciate it if someone points us to the full list.

[1]
https://docs.ceph.com/en/latest/mgr/prometheus/#ceph-daemon-performance-counters-metrics

[2] https://docs.ceph.com/en/latest/monitoring/

-- 
Best regards,
Peter Razumovsky
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io