[ceph-users] Reef release candidate - v18.1.2

2023-06-30 Thread Yuri Weinstein
Hi everyone,

This is the second release candidate for Reef.

The Reef release comes with a new RockDB version (7.9.2) [0], which
incorporates several performance improvements and features. Our
internal testing doesn't show any side effects from the new version,
but we are very eager to hear community feedback on it. This is the
first release to have the ability to tune RockDB settings per column
family [1], which allows for more granular tunings to be applied to
different kinds of data stored in RocksDB. A new set of settings has
been used in Reef to optimize performance for most kinds of workloads
with a slight penalty in some cases, outweighed by large improvements
in use cases such as RGW, in terms of compactions and write
amplification. We would highly encourage community members to give
these a try against their performance benchmarks and use cases. The
detailed list of changes in terms of RockDB and BlueStore can be found
in https://pad.ceph.com/p/reef-rc-relnotes.

If any of our community members would like to help us with performance
investigations or regression testing of the Reef release candidate,
please feel free to provide feedback via email or in
https://pad.ceph.com/p/reef_scale_testing. For more active
discussions, please use the #ceph-at-scale slack channel in
ceph-storage.slack.com.

This RC has gone thru partial testing due to issues we are
experiencing in the sepia lab.
Please try it out and report any issues you encounter. Happy testing!

Thanks,
YuriW
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: [multisite] The purpose of zonegroup

2023-06-30 Thread Casey Bodley
cc Zac, who has been working on multisite docs in
https://tracker.ceph.com/issues/58632

On Fri, Jun 30, 2023 at 11:37 AM Alexander E. Patrakov
 wrote:
>
> Thanks! This is something that should be copy-pasted at the top of
> https://docs.ceph.com/en/latest/radosgw/multisite/
>
> Actually, I reported a documentation bug for something very similar.
>
> On Fri, Jun 30, 2023 at 11:30 PM Casey Bodley  wrote:
> >
> > you're correct that the distinction is between metadata and data;
> > metadata like users and buckets will replicate to all zonegroups,
> > while object data only replicates within a single zonegroup. any given
> > bucket is 'owned' by the zonegroup that creates it (or overridden by
> > the LocationConstraint on creation). requests for data in that bucket
> > sent to other zonegroups should redirect to the zonegroup where it
> > resides
> >
> > the ability to create multiple zonegroups can be useful in cases where
> > you want some isolation for the datasets, but a shared namespace of
> > users and buckets. you may have several connected sites sharing
> > storage, but only require a single backup for purposes of disaster
> > recovery. there it could make sense to create several zonegroups with
> > only two zones each to avoid replicating all objects to all zones
> >
> > in other cases, it could make more sense to isolate things in separate
> > realms with a single zonegroup each. zonegroups just provide some
> > flexibility to control the isolation of data and metadata separately
> >
> > On Thu, Jun 29, 2023 at 5:48 PM Yixin Jin  wrote:
> > >
> > > Hi folks,
> > > In the multisite environment, we can get one realm that contains multiple 
> > > zonegroups, each in turn can have multiple zones. However, the purpose of 
> > > zonegroup isn't clear to me. It seems that when a user is created, its 
> > > metadata is synced to all zones within the same realm, regardless whether 
> > > they are in different zonegroups or not. The same happens to buckets. 
> > > Therefore, what is the purpose of having zonegroups? Wouldn't it be 
> > > easier to just have realm and zones?
> > > Thanks,Yixin
> > > ___
> > > ceph-users mailing list -- ceph-users@ceph.io
> > > To unsubscribe send an email to ceph-users-le...@ceph.io
> > >
> > ___
> > ceph-users mailing list -- ceph-users@ceph.io
> > To unsubscribe send an email to ceph-users-le...@ceph.io
>
>
>
> --
> Alexander E. Patrakov
>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: [multisite] The purpose of zonegroup

2023-06-30 Thread Alexander E. Patrakov
Thanks! This is something that should be copy-pasted at the top of
https://docs.ceph.com/en/latest/radosgw/multisite/

Actually, I reported a documentation bug for something very similar.

On Fri, Jun 30, 2023 at 11:30 PM Casey Bodley  wrote:
>
> you're correct that the distinction is between metadata and data;
> metadata like users and buckets will replicate to all zonegroups,
> while object data only replicates within a single zonegroup. any given
> bucket is 'owned' by the zonegroup that creates it (or overridden by
> the LocationConstraint on creation). requests for data in that bucket
> sent to other zonegroups should redirect to the zonegroup where it
> resides
>
> the ability to create multiple zonegroups can be useful in cases where
> you want some isolation for the datasets, but a shared namespace of
> users and buckets. you may have several connected sites sharing
> storage, but only require a single backup for purposes of disaster
> recovery. there it could make sense to create several zonegroups with
> only two zones each to avoid replicating all objects to all zones
>
> in other cases, it could make more sense to isolate things in separate
> realms with a single zonegroup each. zonegroups just provide some
> flexibility to control the isolation of data and metadata separately
>
> On Thu, Jun 29, 2023 at 5:48 PM Yixin Jin  wrote:
> >
> > Hi folks,
> > In the multisite environment, we can get one realm that contains multiple 
> > zonegroups, each in turn can have multiple zones. However, the purpose of 
> > zonegroup isn't clear to me. It seems that when a user is created, its 
> > metadata is synced to all zones within the same realm, regardless whether 
> > they are in different zonegroups or not. The same happens to buckets. 
> > Therefore, what is the purpose of having zonegroups? Wouldn't it be easier 
> > to just have realm and zones?
> > Thanks,Yixin
> > ___
> > ceph-users mailing list -- ceph-users@ceph.io
> > To unsubscribe send an email to ceph-users-le...@ceph.io
> >
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io



-- 
Alexander E. Patrakov
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: [multisite] The purpose of zonegroup

2023-06-30 Thread Casey Bodley
you're correct that the distinction is between metadata and data;
metadata like users and buckets will replicate to all zonegroups,
while object data only replicates within a single zonegroup. any given
bucket is 'owned' by the zonegroup that creates it (or overridden by
the LocationConstraint on creation). requests for data in that bucket
sent to other zonegroups should redirect to the zonegroup where it
resides

the ability to create multiple zonegroups can be useful in cases where
you want some isolation for the datasets, but a shared namespace of
users and buckets. you may have several connected sites sharing
storage, but only require a single backup for purposes of disaster
recovery. there it could make sense to create several zonegroups with
only two zones each to avoid replicating all objects to all zones

in other cases, it could make more sense to isolate things in separate
realms with a single zonegroup each. zonegroups just provide some
flexibility to control the isolation of data and metadata separately

On Thu, Jun 29, 2023 at 5:48 PM Yixin Jin  wrote:
>
> Hi folks,
> In the multisite environment, we can get one realm that contains multiple 
> zonegroups, each in turn can have multiple zones. However, the purpose of 
> zonegroup isn't clear to me. It seems that when a user is created, its 
> metadata is synced to all zones within the same realm, regardless whether 
> they are in different zonegroups or not. The same happens to buckets. 
> Therefore, what is the purpose of having zonegroups? Wouldn't it be easier to 
> just have realm and zones?
> Thanks,Yixin
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Quincy osd bench in order to define osd_mclock_max_capacity_iops_[hdd|ssd]

2023-06-30 Thread Rafael Diaz Maurin

Hi Luis,

Thank you for sharing your tricks :)

OK it's clever.
You by-pass a destroying disk fio bench with a test on a unique PG on a 
single OSD, and then do some rados bench.

In this way you should get a more realistic ceph IOPS !


Rafael


Le 30/06/2023 à 15:00, Luis Domingues a écrit :

Hi Rafael.

We faced the exact same issue. And we did a bunch of tests and question.

We started with some FIOs, but results where quite meh once in production. Ceph 
bench did not seemed very reliable.

What we ended up doing and seems to hold up quite nicely, is the above. It's 
probably not the best but works for use.

Step 1: We put the disks we want to test in a small dev/test cluster we have. 
So we can mess up with the configs, and go to the not so prod friendly way to 
configure the cluster. And we can make sure no other activity other than our 
test is running.

Step 2: Create a few pools, that have this particularities:
- Only 1 PG
- No replication at all. The only PG of the pool lives in only 1 disk.
- Upmap the PG to the OSD you want to test.
- Repeat for a few disks.

Step 2 needs to enable some options of ceph, as it will not allow you to do it 
by default. I do not remember the exact options, but you can find them on the 
documentation.

Sept 3: set mclock as high_client_ops. So mclock will virtually not limit 
client ops.

Step 4: Run a few ceph rados bench on the different disks.
rados bench  write -t   -b  -p 

 we used 300 secondes to be able do perfom various tests without 
taking a week, but letting the cluster writing a bunch of stuff anyway.
 we used 100, it worked well, and previous tests for other purposes 
showed 100 was nice on our installation.
 the pool with the disk you want to test
 we used 128k and 4M. Feel free to experiment with other values. 
But in our use case it was what we used.

Somewhere in the output of the bench, after it finishes, will be Average IOPS. 
This average is more or less what the disk is capable of handling. Then we put 
some number close to that one. If we have two types of disks that are close to 
each others, we put the smaller value for all disks. And we set it as a global 
configuration on ceph instead of going disk by disk.

It's probably not perfect and looks very like we tinkered something, but it's 
the best solution for testing that so far. And most important, results between 
tests where a lot more consistent than ceph bench or fio.

Hope this will help you.

Luis Domingues
Proton AG


--- Original Message ---
On Friday, June 30th, 2023 at 12:15, Rafael Diaz Maurin 
 wrote:



Hello,

I've just upgraded a Pacific cluster into Quincy, and all my osd have
the low value osd_mclock_max_capacity_iops_hdd : 315.00.

The manuel does not explain how to benchmark the OSD with fio or ceph
bench with good options.
Can someone have the good ceph bench options or fio options in order to
configure osd_mclock_max_capacity_iops_hdd for each osd ?

I ran this bench various times on the same OSD (class hdd) and I obtain
different results.
ceph tell ${osd} cache drop
ceph tell ${osd} bench 12288000 4096 4194304 100

example :
osd.21 (hdd): osd_mclock_max_capacity_iops_hdd = 315.00
bench 1 : 3006.2271379745534
bench 2 : 819.503206458996
bench 3 : 946.5406320134085

How can I succeed in getting the good values for the
osd_mclock_max_capacity_iops_[hdd|ssd] options ?

Thank you for your help,

Rafael





smime.p7s
Description: Signature cryptographique S/MIME
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] db/wal pvmoved ok, but gui show old metadatas

2023-06-30 Thread Christophe BAILLON
Hello,

we have a Ceph 17.2.5 cluster with a total of 26 nodes, where 15 nodes that 
have faulty NVMe drives, 
where the db/wal resides (one NVMe for the first 6 OSDs and another for the 
remaining 6). 

We replaced them with new drives and pvmoved it to avoid losing the OSDs. 

So far, there are no issues, and the OSDs are functioning properly. 

ceph see the correct news disks
root@node02:/# ceph daemon osd.26 list_devices
[
{
"device": "/dev/nvme0n1",
"device_id": "INTEL_SSDPEDME016T4S_CVMD516500851P6KGN"
},
{
"device": "/dev/sdc",
"device_id": "SEAGATE_ST18000NM004J_ZR52TT83C148JFSJ"
}
]

However, the Cephadm GUI still shows the old NVMe drives and hasn't recognized 
the device change.

How can we make the GUI and Cephadm recognize the new devices? 

I tried restarting the managers, thinking that it would rescan the OSDs during 
startup, but it didn't work. 

If you have any ideas, I would appreciate it. 

Should I perform something like that: ceph orch daemon reconfig osd.*

Thank you for your help.
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Quincy osd bench in order to define osd_mclock_max_capacity_iops_[hdd|ssd]

2023-06-30 Thread Luis Domingues
Hi Rafael.

We faced the exact same issue. And we did a bunch of tests and question.

We started with some FIOs, but results where quite meh once in production. Ceph 
bench did not seemed very reliable.

What we ended up doing and seems to hold up quite nicely, is the above. It's 
probably not the best but works for use.

Step 1: We put the disks we want to test in a small dev/test cluster we have. 
So we can mess up with the configs, and go to the not so prod friendly way to 
configure the cluster. And we can make sure no other activity other than our 
test is running.

Step 2: Create a few pools, that have this particularities:
- Only 1 PG
- No replication at all. The only PG of the pool lives in only 1 disk.
- Upmap the PG to the OSD you want to test.
- Repeat for a few disks.

Step 2 needs to enable some options of ceph, as it will not allow you to do it 
by default. I do not remember the exact options, but you can find them on the 
documentation.

Sept 3: set mclock as high_client_ops. So mclock will virtually not limit 
client ops.

Step 4: Run a few ceph rados bench on the different disks.
rados bench  write -t   -b  -p 

 we used 300 secondes to be able do perfom various tests without 
taking a week, but letting the cluster writing a bunch of stuff anyway.
 we used 100, it worked well, and previous tests for other purposes 
showed 100 was nice on our installation.
 the pool with the disk you want to test
 we used 128k and 4M. Feel free to experiment with other values. 
But in our use case it was what we used.

Somewhere in the output of the bench, after it finishes, will be Average IOPS. 
This average is more or less what the disk is capable of handling. Then we put 
some number close to that one. If we have two types of disks that are close to 
each others, we put the smaller value for all disks. And we set it as a global 
configuration on ceph instead of going disk by disk.

It's probably not perfect and looks very like we tinkered something, but it's 
the best solution for testing that so far. And most important, results between 
tests where a lot more consistent than ceph bench or fio.

Hope this will help you.

Luis Domingues
Proton AG


--- Original Message ---
On Friday, June 30th, 2023 at 12:15, Rafael Diaz Maurin 
 wrote:


> Hello,
> 
> I've just upgraded a Pacific cluster into Quincy, and all my osd have
> the low value osd_mclock_max_capacity_iops_hdd : 315.00.
> 
> The manuel does not explain how to benchmark the OSD with fio or ceph
> bench with good options.
> Can someone have the good ceph bench options or fio options in order to
> configure osd_mclock_max_capacity_iops_hdd for each osd ?
> 
> I ran this bench various times on the same OSD (class hdd) and I obtain
> different results.
> ceph tell ${osd} cache drop
> ceph tell ${osd} bench 12288000 4096 4194304 100
> 
> example :
> osd.21 (hdd): osd_mclock_max_capacity_iops_hdd = 315.00
> bench 1 : 3006.2271379745534
> bench 2 : 819.503206458996
> bench 3 : 946.5406320134085
> 
> How can I succeed in getting the good values for the
> osd_mclock_max_capacity_iops_[hdd|ssd] options ?
> 
> Thank you for your help,
> 
> Rafael
> 
> 
> 
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Help needed to configure erasure coding LRC plugin

2023-06-30 Thread Eugen Block

I created a tracker issue, maybe that will get some attention:

https://tracker.ceph.com/issues/61861

Zitat von Michel Jouvin :


Hi Eugen,

Thank you very much for these detailed tests that match what I  
observed and reported earlier. I'm happy to see that we have the  
same understanding of how it should work (based on the  
documentation). Is there any other way that this list to enter in  
contact with the plugin developers as it seems they are not  
following this (very high volume) list... Or may somebody pass the  
email thread to one of them?


Help would be really appreciated. Cheers,

Michel

Le 19/06/2023 à 14:09, Eugen Block a écrit :
Hi, I have a real hardware cluster for testing available now. I'm  
not sure whether I'm completely misunderstanding how it's supposed  
to work or if it's a bug in the LRC plugin.
This cluster has 18 HDD nodes available across 3 rooms (or DCs), I  
intend to use 15 nodes to be able to recover if one node fails.
Given that I need one additional locality chunk per DC I need a  
profile with k + m = 12. So I chose k=9, m=3, l=4 which creates 15  
chunks in total across those 3 DCs, one chunk per host, I checked  
the chunk placement and it is correct. This is the profile I created:


ceph osd erasure-code-profile set lrc1 plugin=lrc k=9 m=3 l=4  
crush-failure-domain=host crush-locality=room crush-device-class=hdd


I created a pool with only one PG to make the output more readable.

This profile should allow the cluster to sustain the loss of three  
chunks, the results are interesting. This is what I tested:


1. I stopped all OSDs on one host and the PG was still active with  
one missing chunk, everything's good.
2. Stopping a second host in the same DC resulted in the PG being  
marked as "down". That was unexpected since with m=3 I expected the  
PG to still be active but degraded. Before test #3 I started all  
OSDs to have the PG active+clean again.
3. I stopped one host per DC, so in total 3 chunks were missing and  
the PG was still active.


Apparently, this profile is able to sustain the loss of m chunks,  
but not an entire DC. I get the impression (and I also discussed  
this with a colleague) that LRC with this implementation is either  
designed to loose only single OSDs which can be recovered quicker  
with fewer surviving OSDs and saving bandwidth. Or this is a bug  
because according to the low-level description [1] the algorithm  
works its way up in the reverse order within the configured layers,  
like in this example (not displaying my k, m, l requirements, just  
for reference):


chunk nr    01234567
step 1  _cDD_cDD
step 2  cDDD
step 3  cDDD

So if a whole DC fails and the chunks from step 3 can not be  
recovered, and maybe step 2 also fails, but eventually step 1  
contains the actual k and m chunks which should sustain the loss of  
an entire DC. My impression is that the algorithm somehow doesn't  
arrive at step 1 and therefore the PG stays down although there are  
enough surviving chunks. I'm not sure if my observations and  
conclusion are correct, I'd love to have a comment from the  
developers on this topic. But in this state I would not recommend  
to use the LRC plugin when the resiliency requirements are to  
sustain the loss of an entire DC.


Thanks,
Eugen

[1]  
https://docs.ceph.com/en/latest/rados/operations/erasure-code-lrc/#low-level-plugin-configuration


Zitat von Michel Jouvin :


Hi,

 I realize that the crushmap I attached to one of my email,  
probably required to understand the discussion here, has been  
stripped down by mailman. To avoid poluting the thread with a long  
output, I put it on at  
https://box.in2p3.fr/index.php/s/J4fcm7orfNE87CX. Download it if  
you are interested.


Best regards,

Michel

Le 21/05/2023 à 16:07, Michel Jouvin a écrit :

Hi Eugen,

My LRC pool is also somewhat experimental so nothing really  
urgent. If you manage to do some tests that help me to understand  
the problem I remain interested. I propose to keep this thread  
for that.


Zitat, I shared my crush map in the email you answered if the  
attachment was not suppressed by mailman.


Cheers,

Michel
Sent from my mobile

Le 18 mai 2023 11:19:35 Eugen Block  a écrit :


Hi, I don’t have a good explanation for this yet, but I’ll soon get
the opportunity to play around with a decommissioned cluster. I’ll try
to get a better understanding of the LRC plugin, but it might take
some time, especially since my vacation is coming up. :-)
I have some thoughts about the down PGs with failure domain OSD, but I
don’t have anything to confirm it yet.

Zitat von Curt :


Hi,

I've been following this thread with interest as it seems like  
a unique use

case to expand my knowledge. I don't use LRC or anything outside basic
erasure coding.

What is your current crush steps rule?  I know you made changes  
since your
first post and had some thoughts I wanted to share, but wanted  
to see your
rule first so I could try 

[ceph-users] Re: ceph-fuse crash

2023-06-30 Thread Milind Changire
If the crash is easily reproducible at your end, could you set debug_client
to 20 in the client-side conf file and then reattempt the operation.

You could then send over the collected logs and we could take a look at
them.

FYI - there's also a bug tracker that has identified a similar problem:
https://tracker.ceph.com/issues/56288, but we don't have the logs for it.



On Fri, Jun 30, 2023 at 11:16 AM  wrote:

> Hi,
>
> I've deployed a ceph-quincy for HPC. Recently, I always encounter the
> problem of ceph-fuse crash
> kernel version is 4.18.0-348.el8.0.2.x86_64
>
> here is part of ceph-fuse log:
>
>-59> 2023-06-28T09:51:00.452+0800 155546ff7700  3 client.159239
> ll_lookup 0x200017f674a.head anaconda3
>-58> 2023-06-28T09:51:00.452+0800 15554cc49700  3 client.159239
> ll_opendir 0x10003e1408d.head
>-57> 2023-06-28T09:51:00.452+0800 15554cc49700  3 client.159239
> may_open 0x1554e79123d0 = 0
>-56> 2023-06-28T09:51:00.452+0800 15554cc49700  3 client.159239
> ll_opendir 0x10003e1408d.head = 0 (0x155328079380)
>-55> 2023-06-28T09:51:00.453+0800 1555473f9700  3 client.159239
> seekdir(0x155328079380, 0)
>-54> 2023-06-28T09:51:00.452+0800 155546bf5700  5 client.159239
> put_cap_ref dropped last FILE_CACHE ref on 0x20004a6e548.head(faked_ino=0
> nref=14 ll_ref=2 cap_refs={1024=0,2048=1} open={1=1} mode=100644
> size=5626/0 nlink=1 btime=2023-06-05T14:38:36.471178+0800
> mtime=2023-06-05T14:38:36.471178+0800 ctime=2023-06-05T14:38:36.471178+0800
> change_attr=1 caps=pAsLsXsFscr(0=pAsLsXsFscr) objectset[0x20004a6e548 ts
> 0/0 objects 1 dirty_or_tx 0] 0x1554e7902300)
>-53> 2023-06-28T09:51:00.453+0800 155546bf5700  3 client.159239 ll_read
> 0x15531806b970 0~8192 = 5626
>-52> 2023-06-28T09:51:00.453+0800 155546ff7700  3 client.159239
> may_lookup 0x155420004840 = 0
>-51> 2023-06-28T09:51:00.453+0800 1554a5dee700  3 client.159239
> ll_lookup 0x10003e1406f.head html
>-50> 2023-06-28T09:51:00.453+0800 155546ff7700  3 client.159239
> ll_lookup 0x200017f674a.head anaconda3 -> 0 (100019b7a43)
>-49> 2023-06-28T09:51:00.453+0800 1554a5dee700  3 client.159239
> may_lookup 0x1554c1a89e40 = 0
>-48> 2023-06-28T09:51:00.453+0800 1554a7dfe700  3 client.159239
> ll_flush 0x15531806b970 0x20004a6e548
>-47> 2023-06-28T09:51:00.453+0800 1555469f4700  3 client.159239
> ll_lookup 0x100019b7a43.head envs
>-46> 2023-06-28T09:51:00.453+0800 1555469f4700  3 client.159239
> may_lookup 0x155420343eb0 = 0
>-45> 2023-06-28T09:51:00.453+0800 1555469f4700  3 client.159239
> ll_lookup 0x100019b7a43.head envs -> 0 (200035de5d8)
>-44> 2023-06-28T09:51:00.453+0800 155545fef700  3 client.159239
> ll_release (fh)0x15531806b970 0x20004a6e548
>-43> 2023-06-28T09:51:00.453+0800 155546bf5700  3 client.159239
> seekdir(0x15544c054010, 1152360438801891331)
>-42> 2023-06-28T09:51:00.453+0800 155546ff7700  3 client.159239
> seekdir(0x15544c029ee0, 1152690945930559491)
>-41> 2023-06-28T09:51:00.453+0800 15554cc49700  3 client.159239
> ll_releasedir 0x15544c029ee0
>-40> 2023-06-28T09:51:00.453+0800 1555452da700  3 client.159239
> ll_lookup 0x20004a66459.head tests
>-39> 2023-06-28T09:51:00.453+0800 15554c244700  3 client.159239
> ll_lookup 0x100040bd04e.head att5410-w -> 0 (200047cbc02)
>-38> 2023-06-28T09:51:00.453+0800 1554a7dfe700  3 client.159239
> ll_lookup 0x200035de5d8.head steven-colossal
>-37> 2023-06-28T09:51:00.453+0800 1554a7dfe700  3 client.159239
> may_lookup 0x1554c20229c0 = 0
>-36> 2023-06-28T09:51:00.453+0800 1554a7dfe700  3 client.159239
> ll_lookup 0x200035de5d8.head steven-colossal -> 0 (100040bcf3a)
>-35> 2023-06-28T09:51:00.453+0800 1555452da700  3 client.159239
> may_lookup 0x155533c970f0 = 0
>-34> 2023-06-28T09:51:00.453+0800 1555452da700  3 client.159239
> ll_lookup 0x20004a66459.head tests -> 0 (20004a6e3e0)
>-33> 2023-06-28T09:51:00.453+0800 155545bed700  3 client.159239
> ll_releasedir 0x15544c054010
>-32> 2023-06-28T09:51:00.453+0800 155546bf5700  3 client.159239
> ll_getattr 0x200047cbc02.head = 0
>-31> 2023-06-28T09:51:00.453+0800 1555452da700  3 client.159239
> ll_lookup 0x200017f674a.head anaconda3
>-30> 2023-06-28T09:51:00.453+0800 1555452da700  3 client.159239
> may_lookup 0x155420004840 = 0
>-29> 2023-06-28T09:51:00.453+0800 1555452da700  3 client.159239
> ll_lookup 0x200017f674a.head anaconda3 -> 0 (100019b7a43)
>-28> 2023-06-28T09:51:00.453+0800 155545fef700  3 client.159239
> ll_lookup 0x100040bcf3a.head lib
>-27> 2023-06-28T09:51:00.453+0800 155545fef700  3 client.159239
> may_lookup 0x1554d7d1d240 = 0
>-26> 2023-06-28T09:51:00.453+0800 155545fef700  3 client.159239
> ll_lookup 0x100040bcf3a.head lib -> 0 (100040bd018)
>-25> 2023-06-28T09:51:00.453+0800 1555469f4700  3 client.159239
> ll_lookup 0x20004a6e3e0.head test_trainer
>-24> 2023-06-28T09:51:00.453+0800 1555469f4700  3 client.159239
> may_lookup 0x155531e5a6c0 = 0
>-23> 2023-06-28T09:51:00.453+0800 15554cc49700  3 

[ceph-users] Quincy osd bench in order to define osd_mclock_max_capacity_iops_[hdd|ssd]

2023-06-30 Thread Rafael Diaz Maurin

Hello,

I've just upgraded a Pacific cluster into Quincy, and all my osd have 
the low value osd_mclock_max_capacity_iops_hdd : 315.00.


The manuel does not explain how to benchmark the OSD with fio or ceph 
bench with good options.
Can someone have the good ceph bench options or fio options in order to 
configure osd_mclock_max_capacity_iops_hdd for each osd ?


I ran this bench various times on the same OSD (class hdd) and I obtain 
different results.

ceph tell ${osd} cache drop
ceph tell ${osd} bench 12288000 4096 4194304 100

example :
osd.21 (hdd): osd_mclock_max_capacity_iops_hdd = 315.00
    bench 1 : 3006.2271379745534
    bench 2 : 819.503206458996
    bench 3 : 946.5406320134085

How can I succeed in getting the good values for the 
osd_mclock_max_capacity_iops_[hdd|ssd] options ?


Thank you for your help,

Rafael





smime.p7s
Description: Signature cryptographique S/MIME
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: RadosGW strange behavior when using a presigned url generated by SDK PHP

2023-06-30 Thread Robin H. Johnson
On Fri, Jun 30, 2023 at 01:21:57AM -, Huy Nguyen wrote:
> Thanks for your reply,
> Yes, my setup is like the following:
> RGWs (port 8084) -> Nginx (80, 443)
> 
> So this why it make me confuse when :8084 appear with the domain.
> 
> And this behavior only occurs with PHP's generated url, not in Boto3
Put tcpdump or something else between nginx & RGW and capture the
transaction when using Boto3 vs PHP.

I'm relatively sure it's nginx adding it for you.

-- 
Robin Hugh Johnson
Gentoo Linux: Dev, Infra Lead, Foundation Treasurer
E-Mail   : robb...@gentoo.org
GnuPG FP : 11ACBA4F 4778E3F6 E4EDF38E B27B944E 34884E85
GnuPG FP : 7D0B3CEB E9B85B1F 825BCECF EE05E6F6 A48F6136


signature.asc
Description: PGP signature
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io