Hi Bill
Am Fr., 5. Dez. 2025 um 10:02 Uhr schrieb Bill Scales <
[email protected]>:
> Hi Reto,
>
> Sorry to hear about your problems with turning on ec optimizations. I've
> led the team that developed this feature, so we are keen to help understand
> what happened to your system.
>
> Your configuration looks fine as far as being supported with ec
> optimizations. The daemons (mons, osds and mgr) need to be running tentacle
> code to use this feature, there is no requirement to update any of your
> clients.
>
> Do you have any more examples of the inconsistent objects logs that you
> can share with me?
>
> There was a bug that we fixed late in the development cycle which was
> scrubbing incorrectly reporting a size mismatch for objects written prior
> to turning on optimizations. This is because prior to turning on
> optimizations objects are padded to a multiple of the stripe_width,
> afterwards objects are no longer padded. The scrub code was occasionally
> getting confused and incorrectly reporting an inconsistency. In this case
> the scrub was reporting false positive errors - there was nothing wrong
> with the data and no problems accessing the data. I notice the log you
> shared in the email was for a size mismatch, I'm interested whether all the
> logs were for mismatched sizes. We will do some further work to confirm
> that the fix is in the 20.2.0 tentacle release and that there are no
> problems with the back port.
>
As far I could see the errors where all about the mismatched sizes.
>
> You also mention that you saw some OSD crashes. Do you have any further
> information about these?
>
I just could cause another crash with starting up a windows VM, and doing a
WIndows file history backup to a drive that is on an RBD Image on pool
rbd_ecpool where I had the allow_ec_optimization flag enabled.
<disk type="network" device="disk">
<driver name="qemu" type="raw" cache="writethrough" io="threads"
discard="unmap" detect_zeroes="unmap"/>
<auth username="admin">
<secret type="ceph" uuid="878b0bc5-c471-4ec6-a92a-f65282ffbdf6"/>
</auth>
<source protocol="rbd" name="rbd/game_windows_backup_drive" index="3">
<host name="zephir" port="6789"/>
</source>
<target dev="vdd" bus="virtio"/>
<alias name="virtio-disk3"/>
<address type="pci" domain="0x0000" bus="0x00" slot="0x02"
function="0x0"/>
</disk>
root@zephir:~# ceph -s
cluster:
id: 27923302-87a5-11ec-ac5b-976d21a49941
health: HEALTH_WARN
2 osds down
Reduced data availability: 49 pgs inactive
Degraded data redundancy: 15722192/79848340 objects degraded
(19.690%), 246 pgs degraded, 247 pgs undersized
services:
mon: 3 daemons, quorum zephir,debian,raspi (age 20h) [leader:
zephir]
mgr: zephir.enywvy(active, since 21h), standbys: debian.nusuye
mds: 3/3 daemons up, 3 standby
osd: 18 osds: 16 up (since 5m), 18 in (since 3d); 8 remapped
pgs
cephfs-mirror: 2 daemons active (2 hosts)
rbd-mirror: 2 daemons active (2 hosts)
rgw: 1 daemon active (1 hosts, 1 zones)
tcmu-runner: 5 portals active (2 hosts)
data:
volumes: 3/3 healthy
pools: 25 pools, 450 pgs
objects: 17.52M objects, 62 TiB
usage: 116 TiB used, 63 TiB / 179 TiB avail
pgs: 10.889% pgs not active
15722192/79848340 objects degraded (19.690%)
197 active+undersized+degraded
183 active+clean
49 undersized+degraded+peered
15 active+clean+scrubbing
5 active+clean+scrubbing+deep
1 active+undersized
io:
client: 1.7 KiB/s rd, 1023 B/s wr, 1 op/s rd, 0 op/s wr
root@zephir:~# ceph osd status
ID HOST USED AVAIL WR OPS WR DATA RD OPS RD DATA STATE
0 debian 11.2T 5289G 0 0 0 0 exists,up
1 zephir 11.6T 4846G 0 0 0 0 exists
2 zephir 10.7T 5827G 0 0 2 4 exists,up
3 zephir 11.0T 5497G 0 0 0 0 exists
4 zephir 159G 1376G 0 0 2 0 exists,up
5 zephir 120G 1415G 0 0 0 0 exists,up
6 debian 11.3T 5200G 0 0 1 0 exists,up
7 zephir 305G 1230G 0 0 0 0 exists,up
8 zephir 10.9T 5594G 0 0 0 0 exists,up
9 zephir 11.9T 4616G 0 0 0 0 exists,up
10 zephir 235G 1300G 0 0 0 0 exists,up
11 zephir 223G 1312G 0 0 0 0 exists,up
12 zephir 11.0T 5479G 0 0 0 0 exists,up
13 debian 118G 631G 0 0 3 8 exists,up
14 debian 345G 1190G 0 0 0 0 exists,up
15 debian 11.4T 5079G 0 0 0 0 exists,up
16 debian 554G 3029G 0 0 3 8 exists,up
17 debian 12.5T 5874G 0 0 1 0 exists,up
root@zephir:~#
I attached the log of osd.1 to
https://filebin.net/jwvs6kuqrc7hx8id
Best Regards,
Reto
>
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]