[ceph-users] Re: Multisite: metadata behind on shards

2024-05-13 Thread Christian Rohmann
On 13.05.24 5:26 AM, Szabo, Istvan (Agoda) wrote: Wonder what is the mechanism behind the sync mechanism because I need to restart all the gateways every 2 days on the remote sites to keep those it in sync. (Octopus 15.2.7) We've also seen lots of those issues with stuck RGWs with earlier

[ceph-users] Re: reef 18.2.3 QE validation status

2024-04-19 Thread Christian Rohmann
On 18.04.24 8:13 PM, Laura Flores wrote: Thanks for bringing this to our attention. The leads have decided that since this PR hasn't been merged to main yet and isn't approved, it will not go in v18.2.3, but it will be prioritized for v18.2.4. I've already added the PR to the v18.2.4 milestone

[ceph-users] Re: reef 18.2.3 QE validation status

2024-04-18 Thread Christian Rohmann
Hey Laura, On 17.04.24 4:58 PM, Laura Flores wrote: There are two PRs that were added later to the 18.2.3 milestone concerning debian packaging: https://github.com/ceph/ceph/pulls?q=is%3Apr+is%3Aopen+milestone%3Av18.2.3 The user is asking if these can be included. I know everybody always

[ceph-users] Re: rgw s3 bucket policies limitations (on users)

2024-04-03 Thread Christian Rohmann
Hey Garcetto, On 29.03.24 4:13 PM, garcetto wrote: i am trying to set bucket policies to allow to different users to access same bucket with different permissions, BUT it seems that is not yet supported, am i wrong? https://docs.ceph.com/en/reef/radosgw/bucketpolicy/#limitations "We do not

[ceph-users] Re: Journal size recommendations

2024-03-08 Thread Christian Rohmann
On 01.03.22 19:57, Eugen Block wrote: can you be more specific what exactly you are looking for? Are you talking about the rocksDB size? And what is the unit for 5012? It’s really not clear to me what you’re asking. And since the recommendations vary between different use cases you might want

[ceph-users] Re: rgw dynamic bucket sharding will hang io

2024-03-08 Thread Christian Rohmann
On 08.03.24 14:25, Christian Rohmann wrote: What do you mean by blocking IO? No bucket actions (read / write) or high IO utilization? According to https://docs.ceph.com/en/latest/radosgw/dynamicresharding/ "Writes to the target bucket are blocked (but reads are not) briefly during resha

[ceph-users] Re: rgw dynamic bucket sharding will hang io

2024-03-08 Thread Christian Rohmann
On 08.03.24 07:22, nuabo tan wrote: When reshard occurs, io will be blocked, why has this serious problem not been solved? Do you care to elaborate on this a bit more? Which Ceph release are you using? Are you using multisite replication or are you talking about a single RGW site? What do

[ceph-users] Re: debian-reef_OLD?

2024-03-05 Thread Christian Rohmann
On 04.03.24 22:24, Daniel Brown wrote: debian-reef/ Now appears to be: debian-reef_OLD/ Could this have been  some sort of "release script" just messing up the renaming / symlinking to the most recent stable? Regards Christian ___ ceph-users

[ceph-users] Re: ceph-crash NOT reporting crashes due to wrong permissions on /var/lib/ceph/crash/posted (Debian / Ubuntu packages)

2024-02-29 Thread Christian Rohmann
On 23.02.24 16:18, Christian Rohmann wrote: I just noticed issues with ceph-crash using the Debian /Ubuntu packages (package: ceph-base): While the /var/lib/ceph/crash/posted folder is created by the package install, it's not properly chowned to ceph:ceph by the postinst script

[ceph-users] ceph-crash NOT reporting crashes due to wrong permissions on /var/lib/ceph/crash/posted (Debian / Ubuntu packages)

2024-02-23 Thread Christian Rohmann
Hey ceph-users, I just noticed issues with ceph-crash using the Debian /Ubuntu packages (package: ceph-base): While the /var/lib/ceph/crash/posted folder is created by the package install, it's not properly chowned to ceph:ceph by the postinst script. This might also affect RPM based

[ceph-users] Re: Throughput metrics missing iwhen updating Ceph Quincy to Reef

2024-02-05 Thread Christian Rohmann
On 01.02.24 10:10, Christian Rohmann wrote: [...] I am wondering if ceph-exporter ([2] is also built and packaged via the ceph packages [3] for installations that use them? [2] https://github.com/ceph/ceph/tree/main/src/exporter [3] https://docs.ceph.com/en/latest/install/get-packages/ I

[ceph-users] Re: how can install latest dev release?

2024-02-01 Thread Christian Rohmann
On 31.01.24 11:33, garcetto wrote: thank you, but seems related to quincy, there is nothing on latest vesions in the doc...maybe the doc is not updated? I don't understand what you are missing. I just used a documentation link pointing to the Quincy version of this page, yes. The "latest"

[ceph-users] Re: Throughput metrics missing iwhen updating Ceph Quincy to Reef

2024-02-01 Thread Christian Rohmann
This change is documented at https://docs.ceph.com/en/latest/mgr/prometheus/#ceph-daemon-performance-counters-metrics, also mentioning the deployment of ceph-exporter which is now used to collect per-host metrics from the local daemons. While this deployment is done by cephadm if used, I am

[ceph-users] Re: how can install latest dev release?

2024-01-31 Thread Christian Rohmann
On 31.01.24 09:38, garcetto wrote: how can i install latest dev release using cephadm? I suppose you found https://docs.ceph.com/en/quincy/install/get-packages/#ceph-development-packages, but yes, that only seems to target a package installation. Would be nice if there were also dev

[ceph-users] Re: RGW rate-limiting or anti-hammering for (external) auth requests // Anti-DoS measures

2024-01-12 Thread Christian Rohmann
Hey Istvan, On 10.01.24 03:27, Szabo, Istvan (Agoda) wrote: I'm using in the frontend https config on haproxy like this, it works so far good: stick-table type ip size 1m expire 10s store http_req_rate(10s) tcp-request inspect-delay 10s tcp-request content track-sc0 src http-request deny

[ceph-users] Re: RGW rate-limiting or anti-hammering for (external) auth requests // Anti-DoS measures

2024-01-09 Thread Christian Rohmann
Happy New Year Ceph-Users! With the holidays and people likely being away, I take the liberty to bluntly BUMP this question about protecting RGW from DoS below: On 22.12.23 10:24, Christian Rohmann wrote: Hey Ceph-Users, RGW does have options [1] to rate limit ops or bandwidth per bucket

[ceph-users] RGW rate-limiting or anti-hammering for (external) auth requests // Anti-DoS measures

2023-12-22 Thread Christian Rohmann
Hey Ceph-Users, RGW does have options [1] to rate limit ops or bandwidth per bucket or user. But those only come into play when the request is authenticated. I'd like to also protect the authentication subsystem from malicious or invalid requests. So in case e.g. some EC2 credentials are not

[ceph-users] Automatic triggering of the Ubuntu SRU process, e.g. for the recent 17.2.7 Quincy point release?

2023-11-12 Thread Christian Rohmann
Hey Yuri, hey ceph-users, first of all, thanks for all your work on developing and maintaining Ceph. I was just wondering if there was any sort of process or trigger to the Ubuntu release team following a point release, for them to also create updated packages. If you look at

[ceph-users] Packages for 17.2.7 released without release notes / announcement (Re: Re: Status of Quincy 17.2.5 ?)

2023-10-30 Thread Christian Rohmann
Sorry to dig up this old thread ... On 25.01.23 10:26, Christian Rohmann wrote: On 20/10/2022 10:12, Christian Rohmann wrote: 1) May I bring up again my remarks about the timing: On 19/10/2022 11:46, Christian Rohmann wrote: I believe the upload of a new release to the repo prior

[ceph-users] CVE-2023-43040 - Improperly verified POST keys in Ceph RGW?

2023-09-27 Thread Christian Rohmann
Hey Ceph-users, i just noticed there is a post to oss-security (https://www.openwall.com/lists/oss-security/2023/09/26/10) about a security issue with Ceph RGW. Signed by IBM / Redhat and including a patch by DO. I also raised an issue on the tracker (https://tracker.ceph.com/issues/63004)

[ceph-users] What is causing *.rgw.log pool to fill up / not be expired (Re: RGW multisite logs (data, md, bilog) not being trimmed automatically?)

2023-09-14 Thread Christian Rohmann
I am unfortunately still observing this issue of the RADOS pool "*.rgw.log" filling up with more and more objects: On 26.06.23 18:18, Christian Rohmann wrote: On the primary cluster I am observing an ever growing (objects and bytes) "sitea.rgw.log" pool, not so on the r

[ceph-users] Re: Can ceph-volume manage the LVs optionally used for DB / WAL at all?

2023-08-26 Thread Christian Rohmann
On 25.08.23 09:09, Eugen Block wrote: I'm still not sure if we're on the same page. Maybe not, I'll respond inline to clarify. By looking at https://docs.ceph.com/en/latest/man/8/ceph-volume/#cmdoption-ceph-volume-lvm-prepare-block.db it seems that ceph-volume wants an LV or partition.

[ceph-users] Re: Can ceph-volume manage the LVs optionally used for DB / WAL at all?

2023-08-25 Thread Christian Rohmann
On 11.08.23 16:06, Eugen Block wrote: if you deploy OSDs from scratch you don't have to create LVs manually, that is handled entirely by ceph-volume (for example on cephadm based clusters you only provide a drivegroup definition). By looking at

[ceph-users] When to use the auth profiles simple-rados-client and profile simple-rados-client-with-blocklist?

2023-08-22 Thread Christian Rohmann
Hey ceph-users, 1) When configuring Gnocchi to use Ceph storage (see https://gnocchi.osci.io/install.html#ceph-requirements) I was wondering if one could use any of the auth profiles like  * simple-rados-client  * simple-rados-client-with-blocklist ? Or are those for different use cases? 2)

[ceph-users] Can ceph-volume manage the LVs optionally used for DB / WAL at all?

2023-08-11 Thread Christian Rohmann
Hey ceph-users, I was wondering if ceph-volume did anything in regards to the management (creation, setting metadata, ) of LVs which are used for DB / WAL of an OSD? Reading the documentation at https://docs.ceph.com/en/latest/man/8/ceph-volume/#new-db is seems to indicate that the LV

[ceph-users] Re: ceph-volume lvm new-db fails

2023-08-11 Thread Christian Rohmann
On 10/08/2023 13:30, Christian Rohmann wrote: It's already fixed master, but the backports are all still pending ... There are PRs for the backports now: * https://tracker.ceph.com/issues/62060 * https://tracker.ceph.com/issues/62061 * https://tracker.ceph.com/issues/62062 Regards

[ceph-users] Re: ceph-volume lvm new-db fails

2023-08-10 Thread Christian Rohmann
On 11/05/2022 23:21, Joost Nieuwenhuijse wrote: After a reboot the OSD turned out to be corrupt. Not sure if ceph-volume lvm new-db caused the problem, or failed because of another problem. I just ran into the same issue trying to add a db to an existing OSD. Apparently this is a known

[ceph-users] Re: RGW accessing real source IP address of a client (e.g. in S3 bucket policies)

2023-07-06 Thread Christian Rohmann
Hey Casey, all, On 16/06/2023 17:00, Casey Bodley wrote: But when applying a bucket policy with aws:SourceIp it seems to only work if I set the internal IP of the HAProxy instance, not the public IP of the client. So the actual remote address is NOT used in my case. Did I miss any config

[ceph-users] Re: RGW multisite logs (data, md, bilog) not being trimmed automatically?

2023-06-29 Thread Christian Rohmann
There was a similar issue reported at https://tracker.ceph.com/issues/48103 and yet another ML post at https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/thread/5LGXQINAJBIGFUZP5WEINVHNPBJEV5X7 May I second the question if it's safe to run radosgw-admin autotrim on those logs? If so, why

[ceph-users] Re: Bluestore compression - Which algo to choose? Zstd really still that bad?

2023-06-27 Thread Christian Rohmann
Hey Igor, On 27/06/2023 12:06, Igor Fedotov wrote: I can't say anything about your primary question on zstd benefits/drawbacks but I'd like to emphasize that compression ratio at BlueStore is (to a major degree) determined by the input data flow characteristics (primarily write block size),

[ceph-users] RGW multisite logs (data, md, bilog) not being trimmed automatically?

2023-06-26 Thread Christian Rohmann
Hey ceph-users, I am running two (now) Quincy clusters doing RGW multi-site replication with only one actually being written to by clients. The other site is intended simply as a remote copy. On the primary cluster I am observing an ever growing (objects and bytes) "sitea.rgw.log" pool, not

[ceph-users] Re: Radogw ignoring HTTP_X_FORWARDED_FOR header

2023-06-26 Thread Christian Rohmann
Hello Yosr, On 26/06/2023 11:41, Yosr Kchaou wrote: We are facing an issue with getting the right value for the header HTTP_X_FORWARDED_FOR when getting client requests. We need this value to do the source ip check validation. [...] Currently, RGW sees that all requests come from 127.0.0.1.

[ceph-users] Bluestore compression - Which algo to choose? Zstd really still that bad?

2023-06-26 Thread Christian Rohmann
Hey ceph-users, we've been using the default "snappy" to have Ceph compress data on certain pools - namely backups / copies of volumes of a VM environment. So it's write once, and no random access. I am now wondering if switching to another algo (there is snappy, zlib, lz4, or zstd) would

[ceph-users] Re: RGW accessing real source IP address of a client (e.g. in S3 bucket policies)

2023-06-16 Thread Christian Rohmann
On 15/06/2023 15:46, Casey Bodley wrote: * In case of HTTP via headers like "X-Forwarded-For". This is apparently supported only for logging the source in the "rgw ops log" ([1])? Or is this info used also when evaluating the source IP condition within a bucket policy? yes, the aws:SourceIp

[ceph-users] RGW accessing real source IP address of a client (e.g. in S3 bucket policies)

2023-06-15 Thread Christian Rohmann
Hello Ceph-Users, context or motivation of my question is S3 bucket policies and other cases using the source IP address as condition. I was wondering if and how RadosGW is able to access the source IP address of clients if receiving their connections via a loadbalancer / reverse proxy like

[ceph-users] Re: pg_autoscaler using uncompressed bytes as pool current total_bytes triggering false POOL_TARGET_SIZE_BYTES_OVERCOMMITTED warnings?

2023-04-21 Thread Christian Rohmann
enlighten me. Thank you and with kind regards Christian On 02/02/2022 20:10, Christian Rohmann wrote: Hey ceph-users, I am debugging a mgr pg_autoscaler WARN which states a target_size_bytes on a pool would overcommit the available storage. There is only one pool with value

[ceph-users] External Auth (AssumeRoleWithWebIdentity) , STS by default, generic policies and isolation by ownership

2023-03-15 Thread Christian Rohmann
Hello ceph-users, unhappy with the capabilities in regards to bucket access policies when using the Keystone authentication module I posted to this ML a while back - https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/message/S2TV7GVFJTWPYA6NVRXDL2JXYUIQGMIN/ In general I'd still like

[ceph-users] Re: Status of Quincy 17.2.5 ?

2023-01-25 Thread Christian Rohmann
Hey everyone, On 20/10/2022 10:12, Christian Rohmann wrote: 1) May I bring up again my remarks about the timing: On 19/10/2022 11:46, Christian Rohmann wrote: I believe the upload of a new release to the repo prior to the announcement happens quite regularly - it might just be due

[ceph-users] Re: OSD slow ops warning not clearing after OSD down

2023-01-16 Thread Christian Rohmann
Hello, On 04/05/2021 09:49, Frank Schilder wrote: I created a ticket: https://tracker.ceph.com/issues/50637 We just observed this very issue on Pacific (16.2.10) , which I also commented on the ticket. I wonder if this case is so seldom, first having some issues causing slow ops and then a

[ceph-users] Re: 16.2.11 branch

2022-12-15 Thread Christian Rohmann
On 15/12/2022 10:31, Christian Rohmann wrote: May I kindly ask for an update on how things are progressing? Mostly I am interested on the (persisting) implications for testing new point releases (e.g. 16.2.11) with more and more bugfixes in them. I guess I just have not looked on the right

[ceph-users] Re: 16.2.11 branch

2022-12-15 Thread Christian Rohmann
Hey Laura, Greg, all, On 31/10/2022 17:15, Gregory Farnum wrote: If you don't mind me asking Laura, have those issues regarding the testing lab been resolved yet? There are currently a lot of folks working to fix the testing lab issues. Essentially, disk corruption affected our ability to

[ceph-users] RGW Forcing buckets to be encrypted (SSE-S3) by default (via a global bucket encryption policy)?

2022-11-23 Thread Christian Rohmann
Hey ceph-users, loosely related to my question about client-side encryption in the Cloud Sync module (https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/thread/I366AIAGWGXG3YQZXP6GDQT4ZX2Y6BXM/) I am wondering if there are other options to ensure data is encrypted at rest and also

[ceph-users] Re: Cloud sync to minio fails after creating the bucket

2022-11-21 Thread Christian Rohmann
On 21/11/2022 12:50, ma...@roterruler.de wrote: Could this "just" be the bug https://tracker.ceph.com/issues/55310 (duplicate https://tracker.ceph.com/issues/57807) about Cloud Sync being broken since Pacific? Wow - yes, the issue seems to be exactly the same that I'm facing -.- But there is

[ceph-users] Re: Cloud sync to minio fails after creating the bucket

2022-11-21 Thread Christian Rohmann
On 21/11/2022 11:04, ma...@roterruler.de wrote: Hi list, I'm currently implementing a sync between ceph and a minio cluster to continously sync the buckets and objects to an offsite location. I followed the guide on https://croit.io/blog/setting-up-ceph-cloud-sync-module After the sync

[ceph-users] Re: RGW replication and multiple endpoints

2022-11-14 Thread Christian Rohmann
Hey Kamil On 14/11/2022 13:54, Kamil Madac wrote: Hello, I'm trying to create a RGW Zonegroup with two zones, and to have data replicated between the zones. Each zone is separate Ceph cluster. There is a possibility to use list of endpoints in zone definitions (not just single endpoint) which

[ceph-users] Re: 16.2.11 branch

2022-10-28 Thread Christian Rohmann
On 28/10/2022 00:25, Laura Flores wrote: Hi Oleksiy, The Pacific RC has not been declared yet since there have been problems in our upstream testing lab. There is no ETA yet for v16.2.11 for that reason, but the full diff of all the patches that were included will be published to ceph.io when

[ceph-users] Re: Status of Quincy 17.2.5 ?

2022-10-20 Thread Christian Rohmann
this week. Thanks for the info. 1) May I bring up again my remarks about the timing: On 19/10/2022 11:46, Christian Rohmann wrote: I believe the upload of a new release to the repo prior to the announcement happens quite regularly - it might just be due to the technical process of releasing. But I

[ceph-users] Mirror de.ceph.com broken?

2022-10-20 Thread Christian Rohmann
Hey ceph-users, it seems that the German ceph mirror http://de.ceph.com/ listed at https://docs.ceph.com/en/latest/install/mirrors/#locations does not hold any data. The index page shows some plesk default page and also deeper links like

[ceph-users] Re: Status of Quincy 17.2.5 ?

2022-10-19 Thread Christian Rohmann
On 19/10/2022 11:26, Chris Palmer wrote: I've noticed that packages for Quincy 17.2.5 appeared in the debian 11 repo a few days ago. However I haven't seen any mention of it anywhere, can't find any release notes, and the documentation still shows 17.2.4 as the latest version. Is 17.2.5

[ceph-users] Re: rgw multisite octopus - bucket can not be resharded after cancelling prior reshard process

2022-10-13 Thread Christian Rohmann
Hey Boris, On 07/10/2022 11:30, Boris Behrens wrote: I just wanted to reshard a bucket but mistyped the amount of shards. In a reflex I hit ctrl-c and waited. It looked like the resharding did not finish so I canceled it, and now the bucket is in this state. How can I fix it. It does not show

[ceph-users] RGW multisite Cloud Sync module with support for client side encryption?

2022-09-12 Thread Christian Rohmann
Hello Ceph-Users, I have a question regarding support for any client side encryption in the Cloud Sync Module for RGW (https://docs.ceph.com/en/latest/radosgw/cloud-sync-module/). While a "regular" multi-site setup (https://docs.ceph.com/en/latest/radosgw/multisite/) is usually syncing

[ceph-users] Re: How to clear "Too many repaired reads on 1 OSDs" on pacific

2022-03-01 Thread Christian Rohmann
On 28/02/2022 20:54, Sascha Vogt wrote: Is there a way to clear the error counter on pacific? If so, how? No, no anymore. See https://tracker.ceph.com/issues/54182 Regards Christian ___ ceph-users mailing list -- ceph-users@ceph.io To

[ceph-users] Re: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools (bug 53663)

2022-02-10 Thread Christian Rohmann
Hey Stefan, thanks for getting back to me! On 10/02/2022 10:05, Stefan Schueffler wrote: since my last mail in Dezember, we changed our ceph-setuo like this: we added one SSD osd on each ceph host (which were pure HDD before). Then, we moved the problematic pool "de-dus5.rgw.buckets.index“

[ceph-users] Re: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools (bug 53663)

2022-02-08 Thread Christian Rohmann
Hey there again, there now was a question from Neha Ojha in https://tracker.ceph.com/issues/53663 about providing OSD debug logs for a manual deep-scrub on (inconsistent) PGs. I did provide the logs of two of those deep-scrubs via ceph-post-file already. But since data inconsistencies are

[ceph-users] Re: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools (bug 53663)

2022-02-07 Thread Christian Rohmann
Hello Ceph-Users! On 22/12/2021 00:38, Stefan Schueffler wrote: The other Problem, regarding the OSD scrub errors, we have this: ceph health detail shows „PG_DAMAGED: Possible data damage: x pgs inconsistent.“ Every now and then new pgs get inconsistent. All inconsistent pgs belong to the

[ceph-users] pg_autoscaler using uncompressed bytes as pool current total_bytes triggering false POOL_TARGET_SIZE_BYTES_OVERCOMMITTED warnings?

2022-02-02 Thread Christian Rohmann
Hey ceph-users, I am debugging a mgr pg_autoscaler WARN which states a target_size_bytes on a pool would overcommit the available storage. There is only one pool with value for  target_size_bytes (=5T) defined and that apparently would consume more than the available storage: --- cut --- #

[ceph-users] Re: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools (bug 53663)

2021-12-21 Thread Christian Rohmann
Thanks for your response Stefan, On 21/12/2021 10:07, Stefan Schueffler wrote: Even without adding a lot of rgw objects (only a few PUTs per minute), we have thousands and thousands of rgw bucket.sync log entries in the rgw log pool (this seems to be a separate problem), and as such we

[ceph-users] Re: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools (bug 53663)

2021-12-21 Thread Christian Rohmann
Hello Eugen, On 20/12/2021 22:02, Eugen Block wrote: you wrote that this cluster was initially installed with Octopus, so no upgrade ceph wise? Are all RGW daemons on the exact same ceph (minor) versions? I remember one of our customers reporting inconsistent objects on a regular basis

[ceph-users] Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools (bug 53663)

2021-12-20 Thread Christian Rohmann
Hello Ceph-Users, for about 3 weeks now I see batches of scrub errors on a 4 node Octopus cluster: # ceph health detail HEALTH_ERR 7 scrub errors; Possible data damage: 6 pgs inconsistent [ERR] OSD_SCRUB_ERRORS: 7 scrub errors [ERR] PG_DAMAGED: Possible data damage: 6 pgs inconsistent    

[ceph-users] Re: inconsistent pg after upgrade nautilus to octopus

2021-12-19 Thread Christian Rohmann
Hello Tomasz, I observe a strange accumulation of inconsistencies for an RGW-only (+multisite) setup, with errors just like those you reported. I collected some info and raised a bug ticket:  https://tracker.ceph.com/issues/53663 Two more inconsistencies have just shown up hours after

[ceph-users] Re: Metrics for object sizes

2021-10-14 Thread Christian Rohmann
On 23/04/2021 03:53, Szabo, Istvan (Agoda) wrote: Objects inside RGW buckets like in couch base software they have their own metrics and has this information. Not as detailed as you would like, but how about using the bucket stats on bucket size and number of objects?  $ radosgw-admin

[ceph-users] Re: Multisite reshard stale instances

2021-10-04 Thread Christian Rohmann
On 04/10/2021 12:22, Christian Rohmann wrote: So there is no reason those instances are still kept? How and when are those instances cleared up? Also just like for the other reporters of this issue, in my case most buckets are deleted buckets, but not all of them. I just hope somebody

[ceph-users] Re: Multisite reshard stale instances

2021-10-04 Thread Christian Rohmann
Hey there again, On 01/10/2021 17:35, Szabo, Istvan (Agoda) wrote: In my setup I've disabled the sharding and preshard each bucket which needs more then 1.1 millions of objects. I also use 11 shards as default, see my ML post

[ceph-users] Re: Multisite reshard stale instances

2021-10-01 Thread Christian Rohmann
On 01/10/2021 17:00, Szabo, Istvan (Agoda) wrote: I just left it and I stopped using synchronous multisite replication. I'm only using directional for a while which is working properly. So you did setup a sync policy to only sync in one direction? In my setup the secondary site does not

[ceph-users] Re: Multisite reshard stale instances

2021-10-01 Thread Christian Rohmann
Hey Istvan, On 05/02/2021 03:00, Szabo, Istvan (Agoda) wrote: I found 6-700 stale instances with the reshard stale instances list command. Is there a way to clean it up (or actually should I clean it up)? The stale instance rm doesn't work in multisite. I observe a similar issue with some

[ceph-users] Re: bucket_index_max_shards vs. no resharding in multisite? How to brace RADOS for huge buckets

2021-09-30 Thread Christian Rohmann
On 30/09/2021 17:02, Christian Rohmann wrote: Looking at my zones I can see that the master zone (converted from previously single-site setup) has  bucket_index_max_shards=0 while the other, secondary zone has  bucket_index_max_shards=11 Should I align this and use "11" as t

[ceph-users] bucket_index_max_shards vs. no resharding in multisite? How to brace RADOS for huge buckets

2021-09-30 Thread Christian Rohmann
Hello Ceph-Users, I just switched from a single to a multi-site setup with all sorts of bucket sizes and large differences in the number of stored objects. Usually resharding is handled by RADOSGW automagically whenever a certain object count per shard is reached, 100k per default. The

[ceph-users] Re: [Suspicious newsletter] Re: create a Multi-zone-group sync setup

2021-08-18 Thread Christian Rohmann
Hey Boris, On 18/08/2021 08:49, Boris Behrens wrote: I've set up realm,first zonegroup with the zone and a sync user in the master setup, and commited. Then I've pulled the periode on the 2nd setup and added a 2nd zonegroup with a zone and commited. Now I can create users in the master setup,

[ceph-users] Re: Multiple DNS names for RGW?

2021-08-18 Thread Christian Rohmann
On 17/08/2021 13:37, Janne Johansson wrote: Don't forget that v4 auth bakes in the clients idea of what the hostname of the endpoint was, so its not only about changing headers. If you are not using v2 auth, you will not be able to rewrite the hostname on the fly. Thanks for the heads up in

[ceph-users] Re: Multiple DNS names for RGW?

2021-08-17 Thread Christian Rohmann
Hey Burkhard, Chris, all, On 16/08/2021 10:48, Chris Palmer wrote: It's straightforward to add multiple DNS names to an endpoint. We do this for the sort of reasons you suggest. You then don't need separate rgw instances (not for this reason anyway). Assuming default:  * radosgw-admin

[ceph-users] Re: pgcalc tool removed (or moved?) from ceph.com ?

2021-07-08 Thread Christian Rohmann
On 08/07/2021 09:39, Dominik Csapak wrote: It's available at https://ceph.com/pgcalc/ just now (with cert not matching), but there apparently are people working on migrating the whole website * ceph.com redirects to https://old.ceph.com/ with matching Let's Encrypt certificate * but

[ceph-users] Re: pgcalc tool removed (or moved?) from ceph.com ?

2021-07-08 Thread Christian Rohmann
Hey Dominik, On 05/07/2021 09:55, Dominik Csapak wrote: Hi, just wanted to ask if it is intentional that http://ceph.com/pgcalc/ results in a 404 error? is there any alternative url? it is still linked from the offical docs. It's available at https://ceph.com/pgcalc/ just now (with cert

[ceph-users] Re: RocksDB degradation / manual compaction vs. snaptrim operations choking Ceph to a halt

2021-07-08 Thread Christian Rohmann
Hey Igor, On 07/07/2021 14:59, Igor Fedotov wrote: after an upgrade from Ceph Nautilus to Octopus we ran into extreme performance issues leading to an unusable cluster when doing a larger snapshot delete and the cluster doing snaptrims, see i.e. https://tracker.ceph.com/issues/50511#note-13.

[ceph-users] Re: rgw multisite sync not syncing data, error: RGW-SYNC:data:init_data_sync_status: ERROR: failed to read remote data log shards

2021-07-07 Thread Christian Rohmann
We found the issue causing data not being synced On 25/06/2021 18:24, Christian Rohmann wrote: What is apparently not working in the sync of actual data. Upon startup the radosgw on the second site shows: 2021-06-25T16:15:06.445+ 7fe71eff5700  1 RGW-SYNC:meta: start 2021-06-25T16:15

[ceph-users] RocksDB degradation / manual compaction vs. snaptrim operations choking Ceph to a halt

2021-07-07 Thread Christian Rohmann
Hello ceph-users, after an upgrade from Ceph Nautilus to Octopus we ran into extreme performance issues leading to an unusable cluster when doing a larger snapshot delete and the cluster doing snaptrims, see i.e. https://tracker.ceph.com/issues/50511#note-13. Since this was not an issue prior

[ceph-users] Re: rgw multisite sync not syncing data, error: RGW-SYNC:data:init_data_sync_status: ERROR: failed to read remote data log shards

2021-06-27 Thread Christian Rohmann
Hey Dominic, thanks for your quick response! On 25/06/2021 19:45, dhils...@performair.com wrote: Christian; Do the second site's RGW instance(s) have access to the first site's OSDs? Is the reverse true? It's been a while since I set up the multi-site sync between our clusters, but I seem

[ceph-users] rgw multisite sync not syncing data, error: RGW-SYNC:data:init_data_sync_status: ERROR: failed to read remote data log shards

2021-06-25 Thread Christian Rohmann
Hey ceph-users, I setup a multisite sync between two freshly setup Octopus clusters. In the first cluster I created a bucket with some data just to test the replication of actual data later. I then followed the instructions on

[ceph-users] RADOSGW Keystone integration - S3 bucket policies targeting not just other tenants / projects ?

2021-06-16 Thread Christian Rohmann
Hallo Ceph-Users, I've been wondering about the state of OpenStack Keystone Auth in RADOSGW. 1) Even though the general documentation on RADOSGW S3 bucket policies is a little "misleading" https://docs.ceph.com/en/latest/radosgw/bucketpolicy/#creation-and-removal in showing users being