David,
We would love some testing of the tool. Are you set up to compile and deploy
Ceph changes? If your situation is not related to the leaked multipart objects
due to retries, it will let you know nothing was fixed. That is still a useful
test.
Another variation of multipart leak comes
Hi,
/var/log/ceph/ceph-osd.0.log is attached.
My sdb is 128MB and sdc (journal) is 16MB:
[root@server0 ~]# ceph-disk list
/dev/dm-0 other, xfs, mounted on /
/dev/dm-1 swap, swap
/dev/sda :
/dev/sda1 other, 0x83
/dev/sda2 other, xfs, mounted on /boot
/dev/sda3 other, LVM2_member
/dev/sdb :
Jewel 10.2.7. I found a discrepancy in object counts for a multisite
configuration and it's looking like it might be orphaned multipart files
causing it. It doesn't look like this PR has received much attention. Is
there anything I can do to help you with testing/confirming a use case for
this
Your bucket index got corrupted. I believe there is no easy way to
restore the index other than downloading existing objects and re-upload
them, correct me if anybody else know a better way.
You can check out all your objects in that bucket with:
rados -p .rgw.buckets ls | grep
Hello!
I've setup LDAP authentication on an object gateway and am attempting to
create a bucket via s3 using python's boto3. It works fine using the access
and secret key for a radosgw user, but access is denied using a token
generated via radosgw-token with the LDAP user's credentials. The user
Should these messages not be gone in 12.2.0?
2017-08-31 20:49:33.500773 7f5aa1756d40 -1 WARNING: the following
dangerous and experimental features are enabled: bluestore
2017-08-31 20:49:33.501026 7f5aa1756d40 -1 WARNING: the following
dangerous and experimental features are enabled: bluestore
All of the messages from sync error list are listed below. The number on
the left is how many times the error message is found.
1811 "message": "failed to sync bucket instance:
(16) Device or resource busy"
7 "message": "failed to sync bucket
Hi Piotr,
Just to verify one point, how are connected your disks (physically), in
a NON-RAID or RAID0 mode ?
rv
Le 31/08/2017 à 16:24, Piotr Dzionek a écrit :
For a last 3 weeks I have been running latest LTS Luminous Ceph
release on CentOS7. It started with 4th RC and now I have Stable
On Thu, 31 Aug 2017, Mark Meyers wrote:
> Hi:
>
> I encountered a metadata lost problem of files when testing the
> outcome of ceph mds's unpredictable breakdown.
> The test shell script is like(not real code):
>
> while i < 10
> do
> touch $i
>
Random theory... I just noticed that the ceph-osd's are listed twice
[1] in the output of systemctl list-dependencies.
Is that correct?!!!
-- dan
[1] > systemctl list-dependencies
...
● ├─ceph-mds.target
● ├─ceph-mon.target
● ├─ceph-osd.target
● │ ├─ceph-osd@48.service
● │ ├─ceph-osd@49.service
Hi:
I encountered a metadata lost problem of files when testing the outcome
of ceph mds's unpredictable breakdown.
The test shell script is like(not real code):
while i < 10
do
touch $i
i++
done
echo c >
Hi,
I see the same with jewel on el7 -- it started one of the recent point
releases around ~10.2.5, IIRC.
Problem seems to be the same -- daemon is started before the osd is
mounted... then the service waits several seconds before trying again.
Aug 31 15:41:47 ceph-osd: 2017-08-31
How long are you seeing these blocked requests for? Initially or
perpetually? Changing the failure domain causes all PGs to peer at the
same time. This would be the cause if it happens really quickly. There is
no way to avoid all of them peering while making a change like this. After
that, It
Datapoint: I have the same issue on 12.1.1, three nodes, 6 disks per node.
On Thu, 31 Aug 2017, Piotr Dzionek said:
> For a last 3 weeks I have been running latest LTS Luminous Ceph release on
> CentOS7. It started with 4th RC and now I have Stable Release.
> Cluster runs fine, however I noticed
Hi Götz,
Something like "rbd status image-spec" usually works for me. Man page says:
"Show the status of the image, including which clients have it open."
I'll tell you which IPs have it open which should help you to track it down.
Cheers,
Maxime
On Thu, 31 Aug 2017 at 16:26 Götz Reinicke
Hi,
Is it possible to see which clients are using an RBD? … I found an RBD in one
of my pools but cant remember if I ever use / mounted it to a client.
Thx for feedback ! Regards . Götz
___
ceph-users mailing list
ceph-users@lists.ceph.com
Hi,
For a last 3 weeks I have been running latest LTS Luminous Ceph release
on CentOS7. It started with 4th RC and now I have Stable Release.
Cluster runs fine, however I noticed that if I do a reboot of one the
nodes, it takes a really long time for cluster to be in ok status.
Osds are
Hi Marcin,
Can you reproduce the crash with 'debug bluestore = 20' set, and then
ceph-post-file /var/log/ceph/ceph-osd.0.log?
My guess is that we're not handling a very small device properly?
sage
On Thu, 31 Aug 2017, Marcin Dulak wrote:
> Hi,
>
> I have a virtual CentOS 7.3 test setup at:
Dear all!
In our Hammer cluster we are planning to switch our failure domain from host to
chassis. We have performed some simulations, and regardless of the settings we
have used some slow requests have appeared all the time.
we had the the following settings:
"osd_max_backfills": "1",
Hi,
Do you know if some of the talks will be in english?
Étienne
From: ceph-users on behalf of Wido den
Hollander
Sent: Thursday, August 24, 2017 17:09
To: ceph-us...@ceph.com
Subject: [ceph-users] Ceph Day
Hi,
I have a virtual CentOS 7.3 test setup at:
https://github.com/marcindulak/github-test-local/blob/a339ff
7505267545f593fd949a6453a56cdfd7fe/vagrant-ceph-rbd-tutorial-centos7.sh
It seems to crash reproducibly with luminous, and works with kraken.
Is this a known issue?
Hi,
Maybe someone have thoughts?
---
Best regards,
Alexander Rudenko
On 30 Aug 2017, at 12:28, Rudenko Aleksandr
> wrote:
Hi,
I use ceph 0.94.10(hammer) with radosgw as S3-compatible object store.
I have few objects in some bucket with strange
22 matches
Mail list logo