[ceph-users] Ceph Pacific 16.2.11 : ceph-volume does not like LV with the same name in different VG

2023-02-06 Thread Gilles Mocellin
Hello ! It seems like ceph-volume from Ceph Pacific 16.2.11 has a problem with same LV names in different VG. I use ceph-ansible (stable-6), with a pre-existing LVM configuration. Here's the error : TASK [ceph-osd : include_tasks scenarios/lvm.yml]

[ceph-users] Re: Removing Rados Gateway in ceph cluster

2023-02-06 Thread Gilles Mocellin
Le 2023-02-06 14:11, Eugen Block a écrit : What does the active mgr log when you try to access the dashboard? Please paste your rgw config settings as well. Ah, Sorry to hijack, but I also can't access Object Storage menus in the Dashboard since upgrading from 16.2.10 to 16.2.11. Here are th

[ceph-users] [RGW] Rebuilding a non master zone

2023-02-28 Thread Gilles Mocellin
Hi Cephers, I have large OMAP objects on one of my cluster (certainly due to a big bucket deletion, and things not completely purged). Since there is no tool to either reconstruct index from data or purge unused index, I thought I can use mutlisite replication. As I am in a multisite config

[ceph-users] Re: Pacific dashboard: unable to get RGW information

2023-04-11 Thread Gilles Mocellin
Hi, My problem is the opposite ! I don't use SSL on RGWs, because I use a load balancer with HTTPS endpoint. so no problem with certificates and IP adresses. With 16.2.11, it does not work anymore because it uses DNS names, and those names are resolving to a management IP, which is not the networ

[ceph-users] Re: Pacific dashboard: unable to get RGW information

2023-04-11 Thread Gilles Mocellin
I forgot, there's a similar bug around that : https://tracker.ceph.com/issues/58811 Le mardi 11 avril 2023, 22:45:28 CEST Gilles Mocellin a écrit : > Hi, > > My problem is the opposite ! > I don't use SSL on RGWs, because I use a load balancer with HTTPS endpoint

[ceph-users] [RGW] Rebuilding a non master zone

2023-04-12 Thread Gilles Mocellin
Hello cephers ! As I was asking in another thread ([RGW] Rebuilding a non master zone), I try to find the best way to rebuild a zone in a multisite config. The goal is to get rid of remaining Large OMAP objects. The simplest way, as I can rely only on the primary zone, is to : - remove the zone

[ceph-users] Re: 17.2.6 Dashboard/RGW Signature Mismatch

2023-04-13 Thread Gilles Mocellin
Le jeudi 13 avril 2023, 18:20:27 CEST Chris Palmer a écrit : > Hi Hello, > I have 3 Ceph clusters, all configured similarly, which have been happy > for some months on 17.2.5: > > 1. A test cluster > 2. A small production cluster > 3. A larger production cluster > > All are debian 11 built

[ceph-users] [RGW] what is log_meta and log_data config in a multisite config?

2023-06-05 Thread Gilles Mocellin
Hi Cephers, In a multisite config, with one zonegroup and 2 zones, when I look at `radiosgw-admin zonegroup get`, I see by defaut these two parameters : "log_meta": "false", "log_data": "true", Where can I find documentation on these, I can't find. I set log_meta to tr

[ceph-users] Re: [RGW] what is log_meta and log_data config in a multisite config?

2023-06-08 Thread Gilles Mocellin
quot; > > I hope that helps. > Rich > > On Mon, 5 Jun 2023 at 20:42, Gilles Mocellin > > wrote: > > Hi Cephers, > > > > In a multisite config, with one zonegroup and 2 zones, when I look at > > `radiosgw-admin zonegroup get`, > > &

[ceph-users] X large objects found in pool 'XXX.rgw.buckets.index'

2023-06-20 Thread Gilles Mocellin
Hello, I still have large OMAP objects since a year. These objects are probably from an ancient bucket that has been removed. So I cannot use bilog trim. Depp-scrub dos nothing. Also, even if I don't have a huge cluster (my Object Storage pools is only arounde 10TB), the rgw-orphan-list is too

[ceph-users] [RGW] Is there a way for a user to change is secret key or create other keys ?

2023-10-09 Thread Gilles Mocellin
Hello Cephers, I was using Ceph with OpenStack, and users could add, remove credentials with `openstack ec2 credentials` commands. But, we are moving our Object Storage service to a new cluster, and didn't want to tie it with OpenStack. Is there a way to have a bit of self service for Rados G

[ceph-users] Re: [RGW] Is there a way for a user to change is secret key or create other keys ?

2023-10-09 Thread Gilles Mocellin
Le lundi 9 octobre 2023, 17:12:53 CEST Casey Bodley a écrit : > On Mon, Oct 9, 2023 at 9:16 AM Gilles Mocellin > > wrote: > > Hello Cephers, > > > > I was using Ceph with OpenStack, and users could add, remove credentials > > with `openstack ec2 credentials` com

[ceph-users] Re: RBD Image Returning 'Unknown Filesystem LVM2_member' On Mount - Help Please

2024-02-04 Thread Gilles Mocellin
Le dimanche 4 février 2024, 09:29:04 CET duluxoz a écrit : > Hi Cedric, > > That's what I thought - the access method shouldn't make a difference. > > No, no lvs details at all - I mean, yes, the osds show up with the lvs > command on the ceph node(s), but not on the individual pools/images (on

[ceph-users] Re: RBD Image Returning 'Unknown Filesystem LVM2_member' On Mount - Help Please

2024-02-04 Thread Gilles Mocellin
Perhaps there is a partition table on your device. What does show : fdisk -l /dev/rbd0 If there is, you can create additional devices with : kpartx /dev/rbd0 And you'll have /dev/rbd0p1 wich will perhaps bee a PV. Le 5 février 2024 07:51:24 GMT+01:00, duluxoz a écrit : >Mounting/Mapping command

[ceph-users] Re: RBD Image Returning 'Unknown Filesystem LVM2_member' On Mount - Help Please

2024-02-07 Thread Gilles Mocellin
Le dimanche 4 février 2024, 09:29:04 CET duluxoz a écrit : > Hi Cedric, > > That's what I thought - the access method shouldn't make a difference. > > No, no lvs details at all - I mean, yes, the osds show up with the lvs > command on the ceph node(s), but not on the individual pools/images (on

[ceph-users] Re: Migrating from S3 to Ceph RGW (Cloud Sync Module)

2024-04-11 Thread Gilles Mocellin
Le jeudi 11 avril 2024, 23:29:30 CEST Casey Bodley a écrit : > unfortunately, this cloud sync module only exports data from ceph to a > remote s3 endpoint, not the other way around: > > "This module syncs zone data to a remote cloud service. The sync is > unidirectional; data is not synced back fr

[ceph-users] Re: Migrating from S3 to Ceph RGW (Cloud Sync Module)

2024-04-11 Thread Gilles Mocellin
Le jeudi 11 avril 2024, 23:44:05 CEST Gilles Mocellin a écrit : > Le jeudi 11 avril 2024, 23:29:30 CEST Casey Bodley a écrit : > > > unfortunately, this cloud sync module only exports data from ceph to a > > remote s3 endpoint, not the other way around: > > > > &qu

[ceph-users] Dashboard : Block image listing and infos

2021-01-18 Thread Gilles Mocellin
Hello Cephers, On a new cluster, I only have 2 RBD block images, and the Dashboard doesn't manage to list them correctly. I have this message : Warning Displaying previously cached data for pool veeam-repos. Sometime it disappears, but as soon as I reload or return to the listing page, it

[ceph-users] Re: Dashboard : Block image listing and infos

2021-01-21 Thread Gilles Mocellin
Hi ! I respond to the list, as it may help others. I also reorder the response. > On Mon, Jan 18, 2021 at 2:41 PM Gilles Mocellin < > > gilles.mocel...@nuagelibre.org> wrote: > > Hello Cephers, > > > > On a new cluster, I only have 2 RBD block images, and th

[ceph-users] XFS block size on RBD / EC vs space amplification

2021-02-02 Thread Gilles Mocellin
Hello, As we know, with 64k for bluestore_min_alloc_size_hdd (I'm only using HDDs), in certain conditions, especially with erasure coding, there's a leak of space while writing objects smaller than 64k x k (EC:k+m). Every object is divided in k elements, written on different OSD. My main us

[ceph-users] Re: XFS block size on RBD / EC vs space amplification

2021-02-03 Thread Gilles Mocellin
5:11 AM Gilles Mocellin wrote: Hello, As we know, with 64k for bluestore_min_alloc_size_hdd (I'm only using HDDs), in certain conditions, especially with erasure coding, there's a leak of space while writing objects smaller than 64k x k (EC:k+m). Every object is divided in k element

[ceph-users] Re: XFS block size on RBD / EC vs space amplification

2021-02-03 Thread Gilles Mocellin
2021, at 13:12, Gilles Mocellin wrote: Hello, As we know, with 64k for bluestore_min_alloc_size_hdd (I'm only using HDDs), in certain conditions, especially with erasure coding, there's a leak of space while writing objects smaller than 64k x k (EC:k+m). Every object is divided

[ceph-users] Re: Dashboard : Block image listing and infos

2021-02-11 Thread Gilles Mocellin
imestamp: Fri Jan 15 18:25:13 2021 access_timestamp: Fri Jan 15 18:25:13 2021 modify_timestamp: Fri Jan 15 19:22:54 2021 Kind Regards,Ernesto On Thu, Jan 21, 2021 at 11:02 PM Gilles Mocellin wrote: Hi ! I respond to the list, as it may help others. I also reorder the respo

[ceph-users] [v15.2.11] radosgw / RGW crash at start, Segmentation Fault

2021-05-07 Thread Gilles Mocellin
Hello, Since I upgrade to Ceph Octopus v15.2.11, on Ubuntu 18.04, Radosgw crash straight at start. On Two clusters, one Lab, and some test on a production cluster, shows the same crash for radosgw. As I don't find any similar bug in the Tracker, neither in this mailing list... Am I alone ?

[ceph-users] Re: [v15.2.11] radosgw / RGW crash at start, Segmentation Fault

2021-05-07 Thread Gilles Mocellin
, so the next point releases should resolve the crashes On Fri, May 7, 2021 at 10:51 AM Gilles Mocellin wrote: Hello, Since I upgrade to Ceph Octopus v15.2.11, on Ubuntu 18.04, Radosgw crash straight at start. On Two clusters, one Lab, and some test on a production cluster, shows the same

[ceph-users] RBD migration between 2 EC pools : very slow

2021-06-22 Thread Gilles Mocellin
Hello Cephers, On a capacitive Ceph cluster (13 nodes, 130 OSDs 8To HDD), I'm migrating a 40 To image from a 3+2 EC pool to a 8+2 one. The use case is Veeam backup on XFS filesystems, mounted via KRBD. Backups are running, and I can see 200MB/s Throughput. But my migration (rbd migrate prep

[ceph-users] Re: RBD migration between 2 EC pools : very slow

2021-06-23 Thread Gilles Mocellin
Le 2021-06-22 20:21, Gilles Mocellin a écrit : Hello Cephers, On a capacitive Ceph cluster (13 nodes, 130 OSDs 8To HDD), I'm migrating a 40 To image from a 3+2 EC pool to a 8+2 one. The use case is Veeam backup on XFS filesystems, mounted via KRBD. Backups are running, and I can see

[ceph-users] Re: RBD migration between 2 EC pools : very slow

2021-06-23 Thread Gilles Mocellin
Le 2021-06-23 14:51, Alexander E. Patrakov a écrit : вт, 22 июн. 2021 г. в 23:22, Gilles Mocellin : Hello Cephers, On a capacitive Ceph cluster (13 nodes, 130 OSDs 8To HDD), I'm migrating a 40 To image from a 3+2 EC pool to a 8+2 one. The use case is Veeam backup on XFS filesy

[ceph-users] How to stop a rbd migration and recover

2021-06-23 Thread Gilles Mocellin
Hello, As a follow-up of the thread "RBD migration between 2 EC pools : very slow". I'm running Octopus 15.2.13. RBD migration seems really fragile. I started a migration to change the data pool (from an EC 3+2 to an EC 8+2) : - rbd migration prepare - rbd migration execute => 4% after 6h, and

[ceph-users] RBD clone to change data pool

2021-07-12 Thread Gilles Mocellin
Hello Cephers, I'm disappointed. I thought I'v had found a good way to migrate from one data pool to another, without too much downtime. I use XFS on RBD, via KRBD to store backups (see another thread). XFS with reflink and crc (accelerate Veeam merges). Also, I want to migrate from a EC k3m2

[ceph-users] Re: RBD clone to change data pool

2021-07-13 Thread Gilles Mocellin
Just some more informations : I use Ceph Octopus 15.2.13 on Ubuntu 18.04, deployed with ceph-ansible, no container (except grafana/prometheus/node_exporter). Le 2021-07-12 23:42, Gilles Mocellin a écrit : Hello Cephers, I'm disappointed. I thought I'v had found a good way to migrat

[ceph-users] Re: Debian 11 Bullseye support

2021-08-27 Thread Gilles Mocellin
Le vendredi 27 août 2021, 09:18:01 CEST Francesco Piraneo G. a écrit : > > For this August Debian testing became a Debian stable with LTS support. > > > > But I see that only sid repo exists, no testing and no new stable > > bullseye. > > > > May be some one knows, when there are plans to have a

[ceph-users] [RGW] Too much index objects and OMAP keys on them

2021-12-01 Thread Gilles Mocellin
Hello, We see large omap objects warnings on the RGW bucket index pool. The objects OMAP keys are about objects in one identified big bucket. Context : = We use S3 storage for an application, with ~1,5 M objects. The production cluster is "replicated" with rclone cron jobs on another d

[ceph-users] [RGW] bi_list(): (5) Input/output error blocking resharding

2022-01-07 Thread Gilles Mocellin
Hello Cephers, Anyone who had that problem find a workaround ? I thing this bug represent well what's all about : https://tracker.ceph.com/issues/51429[1] We have a cluster that it this and cannot auto-reshard. LARGE OMAP Objects stays there and cannot be removed also by pg deep scrub. As

[ceph-users] Re: [RGW] bi_list(): (5) Input/output error blocking resharding

2022-01-10 Thread Gilles Mocellin
Le lundi 10 janvier 2022, 11:42:11 CET Matthew Vernon a ?crit : > Hi, > > On 07/01/2022 18:39, Gilles Mocellin wrote: > > Anyone who had that problem find a workaround ? > > Are you trying to reshard a bucket in a multisite setup? That isn't > expected to work (and,

[ceph-users] Re: [RGW] bi_list(): (5) Input/output error blocking resharding

2022-01-14 Thread Gilles Mocellin
Le 2022-01-11 00:33, Gilles Mocellin a écrit : Le lundi 10 janvier 2022, 11:42:11 CET Matthew Vernon a ?crit : Hi, On 07/01/2022 18:39, Gilles Mocellin wrote: > Anyone who had that problem find a workaround ? Are you trying to reshard a bucket in a multisite setup? That isn't expected

[ceph-users] OSD crash with "no available blob id" / Zombie blobs

2022-03-03 Thread Gilles Mocellin
Hello ! On our Octopus (v15.2.15) cluster, mainly used for OpenStack, We had several OSD crash. Some would not restart with "no available blob id" assertion. We found several related bugs : https://tracker.ceph.com/issues/48216 https://tracker.ceph.com/issues/38272 The workaround that works is

[ceph-users] Re: OSD crash with "no available blob id" / Zombie blobs

2022-03-03 Thread Gilles Mocellin
ter's usage? So major use case is RBD, right? Replicated or EC pools? How often snapshots are taken if any? Thanks, Igor On 3/3/2022 1:45 PM, Gilles Mocellin wrote: Hello ! On our Octopus (v15.2.15) cluster, mainly used for OpenStack, We had several OSD crash. Some would not restart wi

[ceph-users] Replication problems on multi-sites configuration

2022-03-15 Thread Gilles Mocellin
Hello Cephers, We've just configured multi-site between an existing Octopus cluster and a new one in another Datacenter. On our biggest bucket (replic_cfn_prod/cfb : 1.4 M objects, 670 GB) we have many errors like that on the new site : 2022-03-15T10:21:00.800+0100 7f9834750700 1 == st

[ceph-users] Re: Replication problems on multi-sites configuration

2022-03-15 Thread Gilles Mocellin
Hum, it seems I hit that bug : https://tracker.ceph.com/issues/50785 Merged in Pacific, not backported to Octopus, at least not linked to this issue. One more reason to migrate to Pacific... Le 2022-03-15 10:43, Gilles Mocellin a écrit : Hello Cephers, We've just configured multi

[ceph-users] Re: [RGW] Too much index objects and OMAP keys on them

2022-03-26 Thread Gilles Mocellin
rge omap in our RGW index pool... > > On Wed, Dec 1, 2021 at 4:32 AM Gilles Mocellin < > > gilles.mocel...@nuagelibre.org> wrote: > > Hello, > > > > We see large omap objects warnings on the RGW bucket index pool. > > The objects OMAP k

[ceph-users] Re: OSD crash with end_of_buffer + bad crc

2022-04-08 Thread Gilles Mocellin
Some more information about our issue (I work with Wissem). As the OSD are crashing only on one node, we focus on it. We found that it's the only node where we also see taht kind of error in the OSD logs : 2022-04-08T11:38:26.464+0200 7fadaf877700 0 bad crc in data 3052515915 != exp 38845088

[ceph-users] Re: OSD crash with end_of_buffer + bad crc

2022-04-11 Thread Gilles Mocellin
nge the AOC cable (SFP+ + Fibre). But, it's not a Ceph problem. Just a hardware one, that only Ceph has caught ! Le 2022-04-08 11:53, Gilles Mocellin a écrit : Some more information about our issue (I work with Wissem). As the OSD are crashing only on one node, we focus on it. We found th

[ceph-users] Re: OSD crash with end_of_buffer + bad crc

2022-04-26 Thread Gilles Mocellin
Le lundi 11 avril 2022, 10:26:31 CEST Gilles Mocellin a écrit : > Just a follow-up. > > I've found that a specific network interface is causing this. > We have bonds : > - 1 management bond0 > - 1 storage access on bond1 > - 1 storage replication on bond2 > > A

[ceph-users] Re: Bad CRC in data messages logging out to syslog

2022-04-26 Thread Gilles Mocellin
Le lundi 25 avril 2022, 17:46:04 CEST Chris Page a ?crit : > Hi, > > Every now and then I am getting the following logs - > > pve01 2022-04-25T16:41:03.109+0100 7ff35b6da700 0 bad crc in data > 3860390385 != exp 919468086 from v1:10.0.0.111:0/873787122 > pve01 2022-04-25T16:41:04.361+0100 7fb0e2

[ceph-users] Re: OSD crash with end_of_buffer + bad crc

2022-04-27 Thread Gilles Mocellin
Le 2022-04-27 08:43, Konstantin Shalygin a écrit : Just for memo record, what is your network card and driver? ethtool -i eth0 Of course. It's a Dell card, Broadcom Adv. Dual 25Gb Ethernet. $ sudo ethtool -i enp129s0f1np1 driver: bnxt_en version: 1.8.0 firmware-version: 218.0.219.13/pkg 21.8

[ceph-users] RBD images Prometheus metrics : not all pools/images reported

2022-08-16 Thread Gilles Mocellin
Hello Cephers, I'm trying to diagnose who's doing what on our cluster, which suffer from SLOW_OPS, High latency periods since Pacific. And I can't see all pool / images in RBD stats. I had activated RBD image stats while running Octopus, now it seems we only need to define mgr/prometheus/rbd_

[ceph-users] MGR process regularly not responding

2022-10-24 Thread Gilles Mocellin
Hi, In our Ceph Pacific clusters (16.2.10) (1 for OpenStack and S3, 2 for backup on RBD and S3), since the upgrade to Pacific, we have regularly the MGR not responding, not seen anymore in ceph status. The process is still there. Noting in the MGR log, just no more logs. Restarting the servic

[ceph-users] Re: MGR process regularly not responding

2022-10-25 Thread Gilles Mocellin
> just with similar symptoms. > > Regards, > Eugen > > Zitat von Gilles Mocellin : > > Hi, > > > > In our Ceph Pacific clusters (16.2.10) (1 for OpenStack and S3, 2 > > for backup on RBD and S3), > > since the upgrade to Pacific, we have regularly

[ceph-users] OSD failing to restart with "no available blob id"

2020-03-16 Thread Gilles Mocellin
Hi ! I'm stuck with "no available blob id" during the start of an OSD. It seems there's a workaround back-ported only in nautilus (Bug https://tracker.ceph.com/issues/38272), but I use mimic for now. Someone has an operational workaround ? Or should I recreate my OSD ? And what is the easiest way

[ceph-users] Re: Netplan bonding configuration

2020-03-31 Thread Gilles Mocellin
Hello, I don't use netplan, and still on Ubuntu 16.04. But I use VLAN? on the bond, not directly on the interfaces : bond0 : - enp179s0f0 - enp179s0f1 Then I use bond0.323 and bond0.324. (I use a bridge on top to be more like my OpenStack cluster, and with more friendly names : br-mgmt, br-sto

[ceph-users] Re: Netplan bonding configuration

2020-04-01 Thread Gilles Mocellin
parate (2 nics, 1 bond each). >> On 31 Mar 2020, at 18:33, Gilles Mocellin >> wrote: >> >> Hello, >> >> I don't use netplan, and still on Ubuntu 16.04. >> But I use VLAN? on the bond, not directly on the interfaces : >> >> bond0 : >>