[ceph-users] Not timing out watcher

2021-08-07 Thread li jerry
Hello all I use RBD as a data disk for KVM When an abnormal power failure occurs in the kvm server, the rbd watcher often takes 15 minutes to clear automatically. Environmental information: ceph: 15.2.12 os: ubuntu 20.04 kernel: 5.4.0-42-generic libvirt: 6.0.0 QEMU: 4.2.1 Is there any setting t

[ceph-users] Re: rbd object mapping

2021-08-07 Thread Tony Liu
There are two types of "object", RBD-image-object and 8MiB-block-object. When create a RBD image, a RBD-image-object is created and 12800 8MiB-block-objects are allocated. That whole RBD-image-object is mapped to a single PG, which is mapped to 3 OSDs (replica 3). That means, all user data on tha

[ceph-users] Re: rbd object mapping

2021-08-07 Thread Konstantin Shalygin
Object map show where your object with any object name will be placed in defined pool with your crush map, and which of osd will serve this PG. You can type anything in object name - and the the future placement or placement of existing object - this how algo works. 12800 means that your 100GiB

[ceph-users] rbd object mapping

2021-08-07 Thread Tony Liu
Hi, This shows one RBD image is treated as one object, and it's mapped to one PG. "object" here means a RBD image. # ceph osd map vm fcb09c9c-4cd9-44d8-a20b-8961c6eedf8e_disk osdmap e18381 pool 'vm' (4) object 'fcb09c9c-4cd9-44d8-a20b-8961c6eedf8e_disk' -> pg 4.c7a78d40 (4.0) -> up ([4,17,6], p4

[ceph-users] BUG #51821 - client is using insecure global_id reclaim

2021-08-07 Thread Daniel Persson
Hi everyone. I suggested asking for help here instead of in the bug tracker so that I will try it. https://tracker.ceph.com/issues/51821?next_issue_id=51820&prev_issue_id=51824 I have a problem that I can't seem to figure out how to resolve the issue. AUTH_INSECURE_GLOBAL_ID_RECLAIM: client is

[ceph-users] Re: All OSDs on one host down

2021-08-07 Thread Clyso GmbH - Ceph Foundation Member
we have been working and using Ceph adm for more than 2 years.. for this and other reasons have changed our update strategy to immutable infrastructure and are currently in the middle of migrating to different flavours of https://github.com/gardenlinux/gardenlinux. ___

[ceph-users] Re: All OSDs on one host down

2021-08-07 Thread Clyso GmbH - Ceph Foundation Member
Hi Andrew, we have had bad experiences with ubuntu's auto update, especially when updating packages from systemd,dbus and docker. for example: one effect was internal communication errors, only a restart of the node helped. Cheers, Joachim ___ Clyso GmbH - Cep

[ceph-users] Re: All OSDs on one host down

2021-08-07 Thread mabi
Indeed if you upgrade Docker such as with the APT unattended-upgrades the Docker daemon will get restarted meaning all your containers too :( That's just how Docker works. You might want to switch to podman instead of Docker in order to avoid that. I use podman precisely for this reason. ‐

[ceph-users] Re: All OSDs on one host down

2021-08-07 Thread E Taka
A few hours ago we had the same problem, also with Ubuntu 20.04, and there is a coincidence in time with the latest docker update, which was triggered from Puppet. After all, all the containers came back up without a reboot. Thanks for the hint. Note to myself: change the package parameter for the

[ceph-users] Re: All OSDs on one host down

2021-08-07 Thread Andrew Walker-Brown
Yeah I think that’s along the lines of what I’ve faced here. Hopefully i’ve managed to disable the auto updates. Sent from Mail for Windows From: Clyso GmbH - Ceph Foundation Member Sent: 07 August 2021 10:46 T

[ceph-users] Re: All OSDs on one host down

2021-08-07 Thread Andrew Walker-Brown
Thanks David, Spent some more time digging in the logs/google. Also had a further 2 nodes fail this morning (different nodes). Looks like it’s related to apt-auto updates on Ubuntu 20.04, although we don’t run unattended upgrades. Docker appears to get a terminate signal which shutsdown/rest