[ceph-users] Re: Lousy recovery for mclock and reef

2024-05-24 Thread Kai Stian Olstad
_recovery_max_active is default to 3. osd_recovery_op_priority is default to 3, a lower number increases the priority for recovery. All of them can be runtime adjusted. -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe s

[ceph-users] Re: Setting S3 bucket policies with multi-tenants

2024-04-15 Thread Kai Stian Olstad
, "Effect": "Allow", "Principal": { "AWS": [ "arn:aws:iamuser/" ] }, "Action": "s3:*", "Resource": "arn:aws:s3:::*" } ] } And then run s3cmd setpolicy s3-policy.json s3:// -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Large number of misplaced PGs but little backfill going on

2024-03-25 Thread Kai Stian Olstad
On Mon, Mar 25, 2024 at 10:58:24PM +0100, Kai Stian Olstad wrote: On Mon, Mar 25, 2024 at 09:28:01PM +0100, Torkil Svensgaard wrote: My tally came to 412 out of 539 OSDs showing up in a blocked_by list and that is about every OSD with data prior to adding ~100 empty OSDs. How 400 read targets

[ceph-users] Re: Large number of misplaced PGs but little backfill going on

2024-03-25 Thread Kai Stian Olstad
to the new OSD but also between the existing OSD, but they are busy with backfilling so they have no free backfill reservation. I do recommend this document by Joshua Baergen at Digital Ocean that explains backfilling and the problem with it and there solution, a tool called pgremapper. -- Kai Stian

[ceph-users] Re: Large number of misplaced PGs but little backfill going on

2024-03-23 Thread Kai Stian Olstad
On Sat, Mar 23, 2024 at 12:09:29PM +0100, Torkil Svensgaard wrote: The other output is too big for pastebin and I'm not familiar with paste services, any suggestion for a preferred way to share such output? You can attached files to the mail here on the list. -- Kai Stian Olstad

[ceph-users] Re: Reef (18.2): Some PG not scrubbed/deep scrubbed for 1 month

2024-03-22 Thread Kai Stian Olstad
ecause of the 500 limit. So only the SSD got the osd_mclock_max_capacity_iops_ssd set. -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Reef (18.2): Some PG not scrubbed/deep scrubbed for 1 month

2024-03-22 Thread Kai Stian Olstad
osd_mclock_max_capacity_iops_{hdd,ssd} every time the OSD is started. If you check the OSD log you'll see it does the bench. -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: pg repair doesn't fix "got incorrect hash on read" / "candidate had an ec hash mismatch"

2024-03-06 Thread Kai Stian Olstad
, wait for recovery and the run deep-scrub to check if things look better. But would it then use corrupted data on osd 269 to rebuild. - Kai Stian Olstad On 26.02.2024 10:19, Eugen Block wrote: Hi, I think your approach makes sense. But I'm wondering if moving only the problematic PGs

[ceph-users] Re: pg repair doesn't fix "got incorrect hash on read" / "candidate had an ec hash mismatch"

2024-02-27 Thread Kai Stian Olstad
, wait for recovery and the run deep-scrub to check if things look better. But would it then use corrupted data on osd 269 to rebuild. - Kai Stian Olstad On 26.02.2024 10:19, Eugen Block wrote: Hi, I think your approach makes sense. But I'm wondering if moving only the problematic PGs

[ceph-users] Re: pg repair doesn't fix "got incorrect hash on read" / "candidate had an ec hash mismatch"

2024-02-23 Thread Kai Stian Olstad
they are now down from the remaining shards s1(osd.297), s3(osd.276), s4(osd.136) and s5(osd.197) _Any_ comment is highly appreciated. - Kai Stian Olstad On 21.02.2024 13:27, Kai Stian Olstad wrote: Hi, Short summary PG 404.bc is an EC 4+2 where s0 and s2 report hash mismtach for 698 objects

[ceph-users] Re: Some questions about cephadm

2024-02-21 Thread Kai Stian Olstad
. - I am asking because I need to use a local registry to deploy those images. You need to check the cephadm source for the version you would like to use https://github.com/ceph/ceph/blob/v17.2.7/src/cephadm/cephadm#L46 -- Kai Stian Olstad

[ceph-users] pg repair doesn't fix "got incorrect hash on read" / "candidate had an ec hash mismatch"

2024-02-21 Thread Kai Stian Olstad
2.0.46 - test1 [21/Feb/2024:08:27:06.021 +] "GET /benchfiles/2021-11-08T19%3A43%3A50%2C145489260%2B00%3A00 HTTP/1.1" 500 226 - - - latency=0.02568s [1] https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/thread/IPHBE3DLW5ABCZHSNYOBUBSI3TLWVD22/#OE3QXLAJIY6

[ceph-users] Re: PG stuck at recovery

2024-02-19 Thread Kai Stian Olstad
is 4+2 jerasure blaum_roth and the HDD is hybrid where DB is on SSD shared by 5 HDD. And in your cases? -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Installing ceph s3.

2024-02-12 Thread Kai Stian Olstad
can find some documentation on octupus (https://docs.ceph.com/en/octopus/install/ceph-deploy/install-ceph-gateway/) ceph-deploy is deprecated https://docs.ceph.com/en/reef/install/ -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io

[ceph-users] Re: PG stuck at recovery

2024-02-07 Thread Kai Stian Olstad
/IPHBE3DLW5ABCZHSNYOBUBSI3TLWVD22/#OE3QXLAJIY6NU7PNMGHP47UK2CBZJPUG - Kai Stian Olstad On Tue, Feb 06, 2024 at 06:35:26AM -, LeonGao wrote: Hi community We have a new Ceph cluster deployment with 100 nodes. When we are draining an OSD host from the cluster, we see a small amount of PGs that cannot

[ceph-users] Re: how can install latest dev release?

2024-01-31 Thread Kai Stian Olstad
On 31.01.2024 09:38, garcetto wrote: good morning, how can i install latest dev release using cephadm? Have you looked at this page? https://docs.ceph.com/en/latest/install/containers/#development-builds -- Kai Stian Olstad ___ ceph-users mailing

[ceph-users] Re: 17.2.7: Backfilling deadlock / stall / stuck / standstill

2024-01-28 Thread Kai Stian Olstad
cluster. -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: 17.2.7: Backfilling deadlock / stall / stuck / standstill

2024-01-28 Thread Kai Stian Olstad
nk you Wes, switching to wpq and restart the OSDs fixed it for me too. -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] 17.2.7: Backfilling deadlock / stall / stuck / standstill

2024-01-26 Thread Kai Stian Olstad
Hi, This is a cluster running 17.2.7 upgraded from 16.2.6 on the 15 January 2024. On Monday 22 January we had 4 HDD all on different server with I/O-error because of some damage sectors, the OSD is hybrid so the DB is on SSD, 5 HDD share 1 SSD. I set the OSD out, ceph osd out 223 269 290

[ceph-users] Re: podman / docker issues

2024-01-25 Thread Kai Stian Olstad
and specify it, some details are here https://docs.ceph.com/en/reef/cephadm/install/#deployment-in-an-isolated-environment The containers themself don't need to download anything at start. -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io

[ceph-users] Re: Cephadm orchestrator and special label _admin in 17.2.7

2024-01-24 Thread Kai Stian Olstad
and the admin key. -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: About lost disk with erasure code

2023-12-28 Thread Kai Stian Olstad
incy/rados/operations/monitoring-osd-pg/#monitoring-pg-states https://docs.ceph.com/en/quincy/rados/troubleshooting/troubleshooting-pg/ https://docs.ceph.com/en/latest/dev/placement-group/#user-visible-pg-states -- Kai Stian Olstad ___ ceph-users mailing li

[ceph-users] Re: Ceph 16.2.14: osd crash, bdev() _aio_thread got r=-1 ((1) Operation not permitted)

2023-12-03 Thread Kai Stian Olstad
the whole host. I would very much appreciate any suggestions. If you look closely at 53906 you'll see it's a duplicate of https://tracker.ceph.com/issues/53907 In there you have the fix and a workaround until next minor is released. -- Kai Stian Olstad

[ceph-users] Re: ceph osd dump_historic_ops

2023-12-01 Thread Kai Stian Olstad
in install What should I do? The easiest is use tell, then you can run it on any node that have access to ceph. ceph tell osd.8 dump_historic_ops ceph tell osd.8 help will give you all you can do with tell. -- Kai Stian Olstad ___ ceph-users ma

[ceph-users] Re: How to speed up rgw lifecycle

2023-11-28 Thread Kai Stian Olstad
in the documentation, have you tried that? https://docs.ceph.com/en/reef/radosgw/config-ref/#lifecycle-settings -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Ceph 16.2.x excessive logging, how to reduce?

2023-10-09 Thread Kai Stian Olstad
-10-09T07:59:01.303+ This log say that it's bash with PID 4019 that is creating the log entry. Maybe start there, check what what other thing you are running on the server that creates this messages. -- Kai Stian Olstad ___ ceph-users mailing list

[ceph-users] Re: cannot repair a handful of damaged pg's

2023-10-06 Thread Kai Stian Olstad
nice script that do the steps Wes explained above https://github.com/cernceph/ceph-scripts/blob/master/tools/scrubbing/autorepair.sh that you might want to take a look at. -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe s

[ceph-users] Re: Questions about PG auto-scaling and node addition

2023-09-14 Thread Kai Stian Olstad
, but with so large disk you could or maybe should consider having a higher number of PGs per OSD. [1] https://docs.ceph.com/en/quincy/rados/operations/placement-groups/#viewing-pg-scaling-recommendations -- Kai Stian Olstad ___ ceph-users mailing list

[ceph-users] Re: precise/best way to check ssd usage

2023-07-29 Thread Kai Stian Olstad
domain set to host, it's my understanding no data will be migrated/backfilled when a node fails. The reason is that there is nowhere to copy the data to, to fulfill the crush rule one copy on 3 different hosts. -- Kai Stian Olstad ___ ceph-users mailing

[ceph-users] Re: [EXTERNAL] How to change RGW certificate in Cephadm?

2023-06-16 Thread Kai Stian Olstad
mand to do this ceph config-key get mgr/cephadm/spex.rgw.pech | jq -r .spec.spec.rgw_frontend_ssl_certificate | ceph config-key set rgw/cert/rgw.pech - ceph orch restart rgw.pech My claim is that Ceph should update "rgw/cert/rgw.pech" when "mgr/cephadm/spex.rgw.pech" is

[ceph-users] Re: Bottleneck between loadbalancer and rgws

2023-06-15 Thread Kai Stian Olstad
rgw_max_concurrent_requests to 10240 https://www.seagate.com/content/dam/seagate/migrated-assets/www-content/solutions/partners/red-hat/_shared/files/st-seagate-rhcs5-detail-f29951wg-202110-en.pdf So I think the only way to find out it to increase it and see what happens. -- Kai Stian Olstad

[ceph-users] Re: [EXTERNAL] How to change RGW certificate in Cephadm?

2023-06-15 Thread Kai Stian Olstad
PLACEMENT rgw.pech ?:443 7/7 4m ago 22h label:cog But running ceph orch redeploy rgw.pech will redeploy all 7 RGW, and would be the same as ceph orch daemon redeploy rgw.pech.pech-mon-3.upnvrd but only redeploy one of them. From: Kai Stian Olstad The certificate is about

[ceph-users] Re: Bottleneck between loadbalancer and rgws

2023-06-14 Thread Kai Stian Olstad
, but could be increase it to 2048. But the recommended action was to add more gateways instead. [1] https://docs.ceph.com/en/quincy/radosgw/config-ref/#confval-rgw_max_concurrent_requests -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io

[ceph-users] How to change RGW certificate in Cephadm?

2023-06-14 Thread Kai Stian Olstad
ones about to expire. I have looked in the documentation and can't find how to update the certificate for RGW. Can anyone shed some light on how to replace the certificate? -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io

[ceph-users] Re: s3 compatible interface

2023-03-03 Thread Kai Stian Olstad
On Wed, Mar 01, 2023 at 08:39:56AM -0500, Daniel Gryniewicz wrote: We're actually writing this for RGW right now. It'll be a bit before it's productized, but it's in the works. Just curious, what is the use cases for this feature? S3 against CephFS? -- Kai Stian Olstad

[ceph-users] Re: 1 pg recovery_unfound after multiple crash of an OSD

2023-01-09 Thread Kai Stian Olstad
Hi Just a follow up, the issue was solved by running command ceph pg 404.1ff mark_unfound_lost delete - Kai Stian Olstad On 04.01.2023 13:00, Kai Stian Olstad wrote: Hi We are running Ceph 16.2.6 deployed with Cephadm. Around Christmas OSD 245 and 327 had about 20 read error so I set

[ceph-users] 1 pg recovery_unfound after multiple crash of an OSD

2023-01-04 Thread Kai Stian Olstad
oes anyone have any idea on how to resolv the problem? Any help is much appreciated. - Kai Stian Olstad :: ceph-s.txt :: ceph -s --- cluster: id: d13c6b81-51ee-4d22-84e9-456f9307296c health: HEALTH_ERR 1/416287125 objects

[ceph-users] Re: CephFS: Isolating folders for different users

2022-12-23 Thread Kai Stian Olstad
this by looking at how Openstack does it and some trial and error. -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Mails not getting through?

2022-11-16 Thread Kai Stian Olstad
On 16.11.2022 13:21, E Taka wrote: gmail marks too many messages on this mailing list as spam. You can fix that by creating a filter in Gmail for ceph-users@ceph.io and check the "Never send it to Spam". -- Kai Stian Olstad ___ ceph-use

[ceph-users] Re: Mails not getting through?

2022-11-16 Thread Kai Stian Olstad
On 16.11.2022 00:25, Daniel Brunner wrote: are my mails not getting through? is anyone receiving my emails? You can check this yourself by checking the archives https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/ If you see your mail there, they are getting through. -- Kai Stian

[ceph-users] Re: monitoring drives

2022-10-18 Thread Kai Stian Olstad
ort SMART or any status on SAS disk today, I only get the message "No SMART data available". I have gathered "smartctl -x --json=vo" log for the 6 types of SAS this I have in my possession. You can find them here if interested [1] [1] https://gitlab.com/-/s

[ceph-users] Re: Can't setup Basic Ceph Client

2022-07-19 Thread Kai Stian Olstad
eyring? And is the keyring file readable for the user running the ceph commands? -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: cephadm host maintenance

2022-07-14 Thread Kai Stian Olstad
.")[0] }} This works because the first part of the instance name is the hostname. -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Is it normal Ceph reports "Degraded data redundancy" in normal use?

2022-04-19 Thread Kai Stian Olstad
quot; and check status every 2 seconds for 2 minutes, no degraded messages. The only new messages in ceph -s was 12 remapped pgs and "11 active-remapped+backfilling" and "1 active+remapped+backfill_wait" Previously I had to set all osd(15 disks) on a host to out and there was n

[ceph-users] Re: Is it normal Ceph reports "Degraded data redundancy" in normal use?

2022-04-14 Thread Kai Stian Olstad
, but the degradation problems also go away when we turn off the balancer. So this goes a long way of confirming there are something wrong with the balancer since we now see it on two different installation. -- Kai Stian Olstad ___ ceph-users

[ceph-users] Re: ceph namespace access control

2022-03-25 Thread Kai Stian Olstad
it will set the correct permissions. # ceph auth get-or-create client.user2 mon 'profile rbd' osd 'profile rbd pool=rbd namespace=user2' -o /etc/ceph/client.user2.keyring -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: RadosGW S3 range on a 0 byte object gives 416 Range Not Satisfiable

2022-03-22 Thread Kai Stian Olstad
request to me :) Of course, I would never refer someone to send a feature request to a person even if you were a Ceph developer, I would consider that rude, the tracker exist for that :-) -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users

[ceph-users] RadosGW S3 range on a 0 byte object gives 416 Range Not Satisfiable

2022-03-21 Thread Kai Stian Olstad
anyone know it's correct to return 416 on 0 byte object with range or should this be considered a bug in Ceph. [1] https://www.w3.org/Protocols/rfc2616/rfc2616-sec14.html#sec14.35.1 -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users

[ceph-users] Re: Replace HDD with cephadm

2022-03-16 Thread Kai Stian Olstad
On 15.03.2022 10:10, Jimmy Spets wrote: Thanks for your reply. I have two things that I am unsure of: - Is the OSD UUID the same for all OSD:s or should it be unique for each? It's unique and generated when you run ceph-volume lvm prepare or add an OSD. You can find OSD UUID/FSID for

[ceph-users] Re: rbd namespace create - operation not supported

2022-03-16 Thread Kai Stian Olstad
On 11.03.2022 14:04, Ilya Dryomov wrote: On Fri, Mar 11, 2022 at 8:04 AM Kai Stian Olstad wrote: Isn't namespace supported with erasure encoded pools? RBD images can't be created in EC pools, so attempting to create RBD namespaces there is pointless. The way to store RBD image data

[ceph-users] Re: Replace HDD with cephadm

2022-03-11 Thread Kai Stian Olstad
son --osd-fsid 517213f3-0715-4d23-8103-6a34b1f8ef08 And then the OSD should be back up and running. This is the way I have found to do OSD replacement, it might be an easier way of doing it but I have not found that. -- Kai Stian Olstad ___ ceph-users

[ceph-users] rbd namespace create - operation not supported

2022-03-10 Thread Kai Stian Olstad
to add namespace: (95) Operation not supported Isn't namespace supported with erasure encoded pools? -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Unclear on metadata config for new Pacific cluster

2022-02-24 Thread Kai Stian Olstad
> no user data at all. When you have DB on a separate disk the DB size count towards total size of the osd. But this DB space is considered used so you will see a lot of used space. -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@cep

[ceph-users] Re: The Return of Ceph Planet

2022-02-05 Thread Kai Stian Olstad
the change I can't find the feed URL. -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: airgap install

2021-12-21 Thread Kai Stian Olstad
/install/index.html -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: airgap install

2021-12-20 Thread Kai Stian Olstad
sue https://tracker.ceph.com/issues/51667 -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: airgap install

2021-12-16 Thread Kai Stian Olstad
etwork. [1] https://github.com/ceph/ceph/blob/v16.2.7/src/cephadm/cephadm#L50-L61 -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Is it normal Ceph reports "Degraded data redundancy" in normal use?

2021-09-22 Thread Kai Stian Olstad
and is empty except for one pool with one PG created by Cephadm. -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Is it normal Ceph reports "Degraded data redundancy" in normal use?

2021-09-20 Thread Kai Stian Olstad
of the cores was saturated, and network statistics show no errors or drops. So perhaps there is a bug in the balancer code? -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Is it normal Ceph reports "Degraded data redundancy" in normal use?

2021-09-17 Thread Kai Stian Olstad
On 16.09.2021 15:51, Josh Baergen wrote: I assume it's the balancer module. If you write lots of data quickly into the cluster the distribution can vary and the balancer will try to even out the placement. The balancer won't cause degradation, only misplaced objects. Since I'm trying to test

[ceph-users] Is it normal Ceph reports "Degraded data redundancy" in normal use?

2021-09-16 Thread Kai Stian Olstad
data redundancy: 4315/7005845 objects degraded (0.062%), 2 pgs degraded, 4 pgs undersized So my question is, it normal that Ceph report degraded under normal use? or do I have a problem somewhere that I need to investigate? -- Kai Stian Olstad ___ c

[ceph-users] Re: MTU mismatch error in Ceph dashboard

2021-08-06 Thread Kai Stian Olstad
issues/new ? Found a issue reported under MGR https://tracker.ceph.com/issues/52028 - mgr/dashboard: Incorrect MTU mismatch warning -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: MTU mismatch error in Ceph dashboard

2021-08-04 Thread Kai Stian Olstad
a difference but haven’t gotten around to it. If you look at the message it says which interface it is. It does check and report on all the interfaces, even those that is in DOWN state which it shouldn't. -- Kai Stian Olstad ___ ceph-users mailing list

[ceph-users] Re: Cephadm and multipath.

2021-07-29 Thread Kai Stian Olstad
Hi Peter Please remember to include the list address in your reply. I will not trim so people on the list can read you answer. On 29.07.2021 12:43, Peter Childs wrote: On Thu, 29 Jul 2021 at 10:37, Kai Stian Olstad wrote: A little disclaimer, I have never used multipath with Ceph

[ceph-users] Re: Cephadm and multipath.

2021-07-29 Thread Kai Stian Olstad
tipath device paths and not the individual disk the OS sees. [1] https://docs.ceph.com/en/latest/cephadm/osd/#creating-new-osds [2] https://docs.ceph.com/en/latest/cephadm/osd/#advanced-osd-service-specifications -- Kai Stian Olstad ___ ceph-users

[ceph-users] Re: Cephadm: How to remove a stray daemon ghost

2021-07-23 Thread Kai Stian Olstad
On 22.07.2021 13:56, Kai Stian Olstad wrote: Hi I have a warning that says "1 stray daemon(s) not managed by cephadm" What i did is the following. I have 3 nodes that the mon should run on, but because of a bug in 16.2.4 I couldn't run on then since they are in differ

[ceph-users] Cephadm: How to remove a stray daemon ghost

2021-07-22 Thread Kai Stian Olstad
"osd|node|crash" | sort "alertmanager.pech-mds-1" "grafana.pech-cog-2" "mgr.pech-mon-1.ptrsea" "mgr.pech-mon-2.mfdanx" "mon.pech-mon-1" "mon.pech-mon-2" "mon.pech-mon-3" "prometheus.pech-mds-1" No stray daemo

[ceph-users] Re: Having issues to start more than 24 OSDs per host

2021-07-01 Thread Kai Stian Olstad
install/upgrade issue like this has been allowed to sit) You and me both. -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: cephadm: How to replace failed HDD where DB is on SSD

2021-05-27 Thread Kai Stian Olstad
3-8952-eec3e4a2c178 -wi-ao 12.47t osd-block-2da790bc-a74c-41da-8772-3b8aac77001c ceph-block-1b5ad7e7-2e24-4315-8a05-7439ab782b45 -wi-ao 12.47t The fist one is the new OSD and the second one is one that cephadm itself created. -- Kai Stian Olstad ___

[ceph-users] Re: cephadm: How to replace failed HDD where DB is on SSD

2021-05-27 Thread Kai Stian Olstad
on one node. This spec file results in three standalone OSDs! Without the other filters this won't work as expected, it seems. I'll try again on Pacific with the same test and see where that goes. This spec did worked for me when I initially deployed with Octopus 15.2.5. -- Kai Stian Olstad

[ceph-users] Re: cephadm: How to replace failed HDD where DB is on SSD

2021-05-27 Thread Kai Stian Olstad
g volume group backup "/etc/lvm/backup/ceph-block-dbs-563432b7-f52d-4cfe-b952-11542594843b" (seqno 62). INFO:cephadm:/usr/bin/podman:stderr --> Zapping successful for: /dev/ceph-block-dbs-563432b7-f52d-4cfe-b952-11542594843b/osd-block-db-449bd001-eb32-46de-ab80-a1cbcd293d69> -

[ceph-users] Re: cephadm: How to replace failed HDD where DB is on SSD

2021-05-27 Thread Kai Stian Olstad
ocked pech-hd-7 /dev/sds hdd SEAGATE ST14000NM016813.7T No Insufficient space (<10 extents) on vgs, LVM detected, locked -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: cephadm: How to replace failed HDD where DB is on SSD

2021-05-26 Thread Kai Stian Olstad
|DB |WAL | +-+--+-+--++-+ |osd |hdd |pech-hd-7|/dev/sdt |- |-| +-+--+-+--++-+ To me I looks like Cephadm can't use/find the free space on the VG and use that as a new LV for the OSD. -- Kai Stian Olstad ___

[ceph-users] Re: cephadm: How to replace failed HDD where DB is on SSD

2021-05-26 Thread Kai Stian Olstad
tried overwriting the LV with /dev/zero, but that did solve it. So I guess with wipe do you mean delete the LV with lvremove? -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] cephadm: How to replace failed HDD where DB is on SSD

2021-05-25 Thread Kai Stian Olstad
is the wrong way to do it, but I can't find the answer in the documentation. So how can I replace this failed disk in Cephadm? -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Cephadm: Upgrade 15.2.5 -> 15.2.9 stops on non existing OSD

2021-03-11 Thread Kai Stian Olstad
On 11.03.2021 15:47, Sebastian Wagner wrote: yes Am 11.03.21 um 15:46 schrieb Kai Stian Olstad: To resolve it, could I just remove it with "cephadm rm-daemon"? That worked like a charm, and the upgrade is resumed. Thank you Sebastian. -- Kai St

[ceph-users] Re: Cephadm: Upgrade 15.2.5 -> 15.2.9 stops on non existing OSD

2021-03-11 Thread Kai Stian Olstad
ot;stopped", "container_id": null, "container_image_name": "goharbor.example.com/library/ceph/ceph:v15.2.5", "container_image_id": null, "version": null, "started": null, "created": "2021-01-20T09:5

[ceph-users] Cephadm: Upgrade 15.2.5 -> 15.2.9 stops on non existing OSD

2021-03-11 Thread Kai Stian Olstad
at the source code it looks like it get daemons to upgrade from mgr cache, so I restarted both mgr but still it tries to upgrade osd.355. Does anyone know how I can get the upgrade to continue? -- Kai Stian Olstad ___ ceph-users mailing list -- ceph-users