On Monday, June 15, 2015 3:05 AM, ceph-users-requ...@lists.ceph.com
ceph-users-requ...@lists.ceph.com wrote:
Send ceph-users mailing list submissions to
ceph-users@lists.ceph.com
To subscribe or unsubscribe via the World Wide Web, visit
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
or, via email, send a message with subject or body 'help' to
ceph-users-requ...@lists.ceph.com
You can reach the person managing the list at
ceph-users-ow...@lists.ceph.com
When replying, please edit your Subject line so it is more specific
than Re: Contents of ceph-users digest...
Today's Topics:
1. Re: Erasure coded pools and bit-rot protection (Pawe? Sadowski)
2. CephFS client issue (Matteo Dacrema)
3. Re: Erasure coded pools and bit-rot protection (Gregory Farnum)
4. Re: CephFS client issue (Lincoln Bryant)
5. Re: .New Ceph cluster - cannot add additional monitor
(Mike Carlson)
6. Re: CephFS client issue (Matteo Dacrema)
--
Message: 1
Date: Sat, 13 Jun 2015 21:08:25 +0200
From: Pawe? Sadowski c...@sadziu.pl
To: Gregory Farnum g...@gregs42.com
Cc: ceph-users ceph-us...@ceph.com
Subject: Re: [ceph-users] Erasure coded pools and bit-rot protection
Message-ID: 557c7fa9.1020...@sadziu.pl
Content-Type: text/plain; charset=utf-8
Thanks for taking care of this so fast. Yes, I'm getting broken object.
I haven't checked this on other versions but is this bug present
only in Hammer or in all versions?
W dniu 12.06.2015 o 21:43, Gregory Farnum pisze:
Okay, Sam thinks he knows what's going on; here's a ticket:
http://tracker.ceph.com/issues/12000
On Fri, Jun 12, 2015 at 12:32 PM, Gregory Farnum g...@gregs42.com wrote:
On Fri, Jun 12, 2015 at 1:07 AM, Pawe? Sadowski c...@sadziu.pl wrote:
Hi All,
I'm testing erasure coded pools. Is there any protection from bit-rot
errors on object read? If I modify one bit in object part (directly on
OSD) I'm getting *broken*object:
Sorry, are you saying that you're getting a broken object if you flip
a bit in an EC pool? That should detect the chunk as invalid and
reconstruct on read...
-Greg
mon-01:~ # rados --pool ecpool get `hostname -f`_16 - | md5sum
bb2d82bbb95be6b9a039d135cc7a5d0d -
# modify one bit directly on OSD
mon-01:~ # rados --pool ecpool get `hostname -f`_16 - | md5sum
02f04f590010b4b0e6af4741c4097b4f -
# restore bit to original value
mon-01:~ # rados --pool ecpool get `hostname -f`_16 - | md5sum
bb2d82bbb95be6b9a039d135cc7a5d0d -
If I run deep-scrub on modified bit I'm getting inconsistent PG which is
correct in this case. After restoring bit and running deep-scrub again
all PGs are clean.
[ceph version 0.94.1 (e4bfad3a3c51054df7e537a724c8d0bf9be972ff)]
--
PS
--
Message: 2
Date: Sun, 14 Jun 2015 15:26:54 +
From: Matteo Dacrema mdacr...@enter.it
To: ceph-users ceph-us...@ceph.com
Subject: [ceph-users] CephFS client issue
Message-ID: d28e061762104ed68e06effd5199ef06@Exch2013Mb.enter.local
Content-Type: text/plain; charset=us-ascii
?Hi all,
I'm using CephFS on Hammer and sometimes I need to reboot one or more clients
because , as ceph -s tells me, it's failing to respond to capability
release.After tha?t all clients stop to respond: can't access files or
mount/umont cephfs.
I've 1.5 million files , 2 metadata servers in active/standby configuration
with 8 GB of RAM , 20 clients with 2 GB of RAM each and 2 OSD nodes with 4 80GB
osd and 4GB of RAM.
Here my configuration:
[global]
fsid = 2de7b17f-0a3e-4109-b878-c035dd2f7735
mon_initial_members = cephmds01
mon_host = 10.29.81.161
auth_cluster_required = cephx
auth_service_required = cephx
auth_client_required = cephx
public network = 10.29.81.0/24
tcp nodelay = true
tcp rcvbuf = 0
ms tcp read timeout = 600
#Capacity
mon osd full ratio = .95
mon osd nearfull ratio = .85
[osd]
osd journal size = 1024
journal dio = true
journal aio = true
osd op threads = 2
osd op thread timeout = 60
osd disk threads = 2
osd recovery threads = 1
osd recovery max active = 1
osd max backfills = 2
# Pool
osd pool default size = 2
#XFS
osd mkfs type = xfs
osd mkfs options xfs = -f -i size=2048
osd mount options xfs = rw,noatime,inode64,logbsize=256k,delaylog
#FileStore Settings
filestore xattr use omap = false
filestore max inline xattr size = 512
filestore max sync interval = 10
filestore merge threshold = 40
filestore split multiple = 8
filestore flusher = false
filestore queue max ops = 2000
filestore queue max bytes = 536870912
filestore queue committing max ops = 500
filestore queue committing max bytes =