Re: Ceph backports workflow update

2015-02-11 Thread Gregory Farnum
On Wed, Feb 11, 2015 at 10:19 AM, Loic Dachary l...@dachary.org wrote: On 11/02/2015 18:27, Gregory Farnum wrote: Mmm. I'm happy to look at suites that get run this way but I'm unlikely to notice them go by on the list if I'm not poked about them — I generally filter out anything that has

Re: dumpling integration branch for v0.67.12 ready for QE

2015-02-10 Thread Gregory Farnum
On Tue, Feb 10, 2015 at 9:22 AM, Loic Dachary l...@dachary.org wrote: On 10/02/2015 18:19, Yuri Weinstein wrote: Loic, The only difference between options if we run suits on merged dumpling vs dumpling-backports first - is time. We will have to run suites on the final branch after the

Re: dumpling integration branch for v0.67.12 ready for QE

2015-02-10 Thread Gregory Farnum
On Tue, Feb 10, 2015 at 10:04 AM, Loic Dachary l...@dachary.org wrote: On 10/02/2015 18:29, Yuri Weinstein wrote: On 10/02/2015 18:19, Yuri Weinstein wrote: Loic, The only difference between options if we run suits on merged dumpling vs dumpling-backports first - is time. We will have to

Re: K/V interface buffer transaction

2015-02-10 Thread Gregory Farnum
On Tue, Feb 10, 2015 at 10:26 AM, Sage Weil sw...@redhat.com wrote: On Tue, 10 Feb 2015, Somnath Roy wrote: Thanks Sam ! So, is it safe to do ordering if in a transaction *no* remove/truncate/create/add call ? For example, do we need to preserve ordering in case of the below transaction ?

Re: dumpling integration branch for v0.67.12 ready for QE

2015-02-10 Thread Gregory Farnum
On Tue, Feb 10, 2015 at 10:33 AM, Loic Dachary l...@dachary.org wrote: On 10/02/2015 19:25, Gregory Farnum wrote: Now, as it happens there are some reasons to maintain a dumpling branch that isn't part of backports. We've been doing a lot of work lately to make the nightlies behave well

Re: speed decrease since firefly,giant,hammer the 2nd try

2015-02-10 Thread Gregory Farnum
On Tue, Feb 10, 2015 at 10:55 AM, Stefan Priebe s.pri...@profihost.ag wrote: Hello, last year in june i already reported this but there was no real result. (http://lists.ceph.com/pipermail/ceph-users-ceph.com/2014-July/041070.html) I then had the hope that this will be fixed itself when

Re: [ceph-commit] branch master updated. v0.92-935-g541a8f7

2015-02-10 Thread Gregory Farnum
Haomai, On Sun, Feb 8, 2015 at 2:22 AM, ceph.git ceph-com...@ceph.com wrote: This is an automated email from the git hooks/post-receive script. It was generated because a ref change was pushed to the repository containing the project . The branch, master has been updated via

Re: Bucket lifecycle (object expiration)

2015-02-09 Thread Gregory Farnum
On Mon, Feb 9, 2015 at 7:33 AM, Yehuda Sadeh-Weinraub yeh...@redhat.com wrote: - Original Message - From: Sage Weil s...@newdream.net To: Yehuda Sadeh-Weinraub yeh...@redhat.com Cc: Ceph Development ceph-devel@vger.kernel.org Sent: Monday, February 9, 2015 3:42:40 AM Subject: Re:

Re: scrub scheduling

2015-02-09 Thread Gregory Farnum
On Sun, Feb 8, 2015 at 1:38 PM, Sage Weil sw...@redhat.com wrote: Simon Leinen at Switch did a greaet post recently about the impact of scrub on their cluster(s): http://blog.simon.leinen.ch/2015/02/ceph-deep-scrubbing-impact.html Basically the 2 week deep scrub interval kicks in on

Re: Bounding OSD memory requirements during peering/recovery

2015-02-09 Thread Gregory Farnum
Right. So, memory usage of an OSD is usually linear in the number of PGs it hosts. However, that memory can also grow based on at least one other thing: the number of OSD Maps required to go through peering. It *looks* to me like this is what you're running in to, not growth on the number of

Re: About in_seq, out_seq in Messenger

2015-02-09 Thread Gregory Farnum
- Original Message - From: Haomai Wang haomaiw...@gmail.com To: Gregory Farnum gfar...@redhat.com Cc: Sage Weil sw...@redhat.com, ceph-devel@vger.kernel.org Sent: Friday, February 6, 2015 8:16:42 AM Subject: Re: About in_seq, out_seq in Messenger On Fri, Feb 6, 2015 at 10:47 PM

Re: Bucket lifecycle (object expiration)

2015-02-06 Thread Gregory Farnum
On Fri, Feb 6, 2015 at 1:46 PM, Yehuda Sadeh-Weinraub yeh...@redhat.com wrote: I have been recently looking at implementing object expiration in rgw. First, a brief description of the feature: S3 provides mechanisms to expire objects, and/or to transition them into different storage class.

Re: Confused about SnapMapper::get_prefix

2015-02-03 Thread Gregory Farnum
On Tue, Feb 3, 2015 at 4:12 AM, Ding Dinghua dingdinghu...@gmail.com wrote: Hi all: I don't understand why SnapMapper::get_prefix static_cast snap to unsigned: string SnapMapper::get_prefix(snapid_t snap) { char buf[100]; int len = snprintf( buf, sizeof(buf), %.*X_,

Re: [PATCH] crc32c: add aarch64 optimized crc32c implementation

2015-02-03 Thread Gregory Farnum
The guys who would normally handle this have been traveling lately, so they're probably behind on such things. That said, Github pull requests are probably a more reliable transmission channel than the mailing list. :) -Greg On Tue, Feb 3, 2015 at 11:31 AM, Yazen Ghannam yazen.ghan...@linaro.org

Re: workloads/rbd_fsx_cache_writethrough.yaml hangs on dumpling

2015-02-02 Thread Gregory Farnum
On Mon, Feb 2, 2015 at 7:00 AM, Loic Dachary l...@dachary.org wrote: On 02/02/2015 14:48, Yan, Zheng wrote: On Mon, Feb 2, 2015 at 9:18 PM, Loic Dachary l...@dachary.org wrote: Hi, http://pulpito.ceph.com/loic-2015-01-29_15:39:38-rbd-dumpling-backports---basic-multi/730029/ hangs on

Re: [ceph-users] chattr +i not working with cephfs

2015-01-28 Thread Gregory Farnum
On Wed, Jan 28, 2015 at 10:06 AM, Sage Weil s...@newdream.net wrote: On Wed, 28 Jan 2015, John Spray wrote: On Wed, Jan 28, 2015 at 5:23 PM, Gregory Farnum g...@gregs42.com wrote: My concern is whether we as the FS are responsible for doing anything more than storing and returning

Re: [ceph-users] chattr +i not working with cephfs

2015-01-28 Thread Gregory Farnum
On Wed, Jan 28, 2015 at 5:24 AM, John Spray john.sp...@redhat.com wrote: We don't implement the GETFLAGS and SETFLAGS ioctls used for +i. Adding the ioctls is pretty easy, but then we need somewhere to put the flags. Currently we don't store a flags attribute on inodes, but maybe we could

Re: Article about Object Store API

2015-01-27 Thread Gregory Farnum
Nifty; it's good to have that sort of blog-style documentation about the interface. Are you planning to do some work with it that you can show off as well? :) -Greg On Tue, Jan 27, 2015 at 12:48 PM, Marcel Lauhoff m...@irq0.org wrote: Hi, I wrote an article about the object store API - How it

Re: idempotent op (esp delete)

2015-01-23 Thread Gregory Farnum
On Fri, Jan 23, 2015 at 1:43 PM, Sage Weil sw...@redhat.com wrote: Background: 1) Way back when we made a task that would thrash the cache modes by adding and removing the cache tier while ceph_test_rados was running. This mostly worked, but would occasionally fail because we would -

Re: idempotent op (esp delete)

2015-01-23 Thread Gregory Farnum
On Fri, Jan 23, 2015 at 2:18 PM, Sage Weil sw...@redhat.com wrote: On Fri, 23 Jan 2015, Gregory Farnum wrote: On Fri, Jan 23, 2015 at 1:43 PM, Sage Weil sw...@redhat.com wrote: Background: 1) Way back when we made a task that would thrash the cache modes by adding and removing the cache

Re: mon: forwarding user commands

2015-01-19 Thread Gregory Farnum
On Sun, Jan 18, 2015 at 11:02 PM, Mykola Golub mgo...@mirantis.com wrote: On Sun, Jan 18, 2015 at 10:33:05AM -0800, Sage Weil wrote: On Sun, 18 Jan 2015, Mykola Golub wrote: Hi Ceph, Right now, for not a monitor leader, if a received command is not supported locally, but is supported by

Re: New Defects reported by Coverity Scan for ceph

2015-01-16 Thread Gregory Farnum
Sage, are these uncaught assertion errors something we normally ignore? I'm not familiar with any code that tries to catch errors in our standard init patterns, which is what looks to be the problem on these new coverity issues in cephfs-table-tool. -Greg On Fri, Jan 16, 2015 at 6:39 AM,

Re: MDS has inconsistent performance

2015-01-16 Thread Gregory Farnum
On Fri, Jan 16, 2015 at 10:34 AM, Michael Sevilla mikesevil...@gmail.com wrote: On Thu, Jan 15, 2015 at 10:37 PM, Gregory Farnum g...@gregs42.com wrote: On Thu, Jan 15, 2015 at 2:44 PM, Michael Sevilla mikesevil...@gmail.com wrote: Let me know if this works and/or you need anything else

Re: Is there any way to collect object-level I/O access traces in Ceph?

2015-01-16 Thread Gregory Farnum
On Fri, Jan 16, 2015 at 11:28 AM, Lipeng Wan lipengwa...@gmail.com wrote: Dear all, Does Ceph provide a way to collect object-level I/O access traces? Specifically, can we collect the traces to record how many times each object has been accessed (read, write, etc.) during a fixed period of

Re: Is there any way to collect object-level I/O access traces in Ceph?

2015-01-16 Thread Gregory Farnum
, ceph-mds.ceph-node1.log, etc., which log file should I look at? Maybe the ceph-mds.ceph-node1.log? Specifically, is there any keyword I can search in the log file to locate the object operations? Thanks! LW On Fri, Jan 16, 2015 at 4:21 PM, Gregory Farnum g...@gregs42.com wrote: On Fri, Jan

Re: QoS Control for RBD I/Os?

2015-01-15 Thread Gregory Farnum
? Cheng Cheng On Thu, Jan 15, 2015 at 12:59 PM, Gregory Farnum g...@gregs42.com wrote: On Thu, Jan 15, 2015 at 9:53 AM, Cheng Cheng ccheng@gmail.com wrote: Hi Ceph, I am wondering is there a mechanism to prioritize the rbd_aio_write/rbd_aio_read I/Os? Currently all RBD I/Os

Re: QoS Control for RBD I/Os?

2015-01-15 Thread Gregory Farnum
On Thu, Jan 15, 2015 at 9:53 AM, Cheng Cheng ccheng@gmail.com wrote: Hi Ceph, I am wondering is there a mechanism to prioritize the rbd_aio_write/rbd_aio_read I/Os? Currently all RBD I/Os are issued in FIFO to rados layer, and there is NO QoS mechanism to control the priority of these

Re: 'Immutable bit' on pools to prevent deletion

2015-01-15 Thread Gregory Farnum
On Thu, Jan 15, 2015 at 9:44 AM, Sage Weil sw...@redhat.com wrote: In addition (or instead of) making the API harder to fat-finger, we could also add a mon config option like mon allow pool deletion = false that defaults off. Then, to delete any pool, you need to update ceph.conf and

Re: 'Immutable bit' on pools to prevent deletion

2015-01-15 Thread Gregory Farnum
Oh, I think it might be fine to require setting a config option before you delete stuff; I just don't want to prevent the option from being injectable. :) On Thu, Jan 15, 2015 at 10:07 AM, Sage Weil sw...@redhat.com wrote: On Thu, 15 Jan 2015, Gregory Farnum wrote: On Thu, Jan 15, 2015 at 9:44

Re: MDS has inconsistent performance

2015-01-15 Thread Gregory Farnum
On Thu, Jan 15, 2015 at 2:44 PM, Michael Sevilla mikesevil...@gmail.com wrote: Let me know if this works and/or you need anything else: https://www.dropbox.com/s/fq47w6jebnyluu0/lookup-logs.tar.gz?dl=0 Beware - the clients were on debug=10. Also, I tried this with the kernel client and it is

Re: MDS has inconsistent performance

2015-01-15 Thread Gregory Farnum
Can you post the full logs somewhere to look at? These bits aren't very helpful on their own (except to say, yes, the client cleared its I_COMPLETE for some reason). On Tue, Jan 13, 2015 at 3:45 PM, Michael Sevilla mikesevil...@gmail.com wrote: On Tue, Jan 13, 2015 at 11:13 AM, Gregory Farnum g

Re: MDS has inconsistent performance

2015-01-13 Thread Gregory Farnum
On Mon, Jan 12, 2015 at 10:17 PM, Michael Sevilla mikesevil...@gmail.com wrote: I can't get consistent performance with 1 MDS. I have 2 clients create 100,000 files (separate directories) in a CephFS mount. I ran the experiment 5 times (deleting the pools/fs and restarting the MDS in between

Re: /usr/bin/cephfs tool

2015-01-09 Thread Gregory Farnum
On Fri, Jan 9, 2015 at 7:20 AM, Sage Weil sw...@redhat.com wrote: Should we drop this entirely in hammer? Yes! If I remember correctly all of the layout stuff is fully supported using virtual xattrs and standard tools. The only thing left is the tool that shows you how file blocks map to

Re: /usr/bin/cephfs tool

2015-01-09 Thread Gregory Farnum
On Fri, Jan 9, 2015 at 10:00 AM, Sage Weil sw...@redhat.com wrote: On Fri, 9 Jan 2015, Gregory Farnum wrote: On Fri, Jan 9, 2015 at 7:20 AM, Sage Weil sw...@redhat.com wrote: Should we drop this entirely in hammer? Yes! If I remember correctly all of the layout stuff is fully supported

Re: /usr/bin/cephfs tool

2015-01-09 Thread Gregory Farnum
On Fri, Jan 9, 2015 at 10:08 AM, Sage Weil sw...@redhat.com wrote: On Fri, 9 Jan 2015, Gregory Farnum wrote: On Fri, Jan 9, 2015 at 10:00 AM, Sage Weil sw...@redhat.com wrote: On Fri, 9 Jan 2015, Gregory Farnum wrote: On Fri, Jan 9, 2015 at 7:20 AM, Sage Weil sw...@redhat.com wrote

Re: Ceph backports

2015-01-06 Thread Gregory Farnum
On Tue, Jan 6, 2015 at 12:39 AM, Loic Dachary l...@dachary.org wrote: On 06/01/2015 01:22, Gregory Farnum wrote: On Mon, Jan 5, 2015 at 4:12 PM, Loic Dachary l...@dachary.org wrote: :-) This process is helpful if it allows me to help a little more than I currently do with the backport

deprecated rbd functions and warnings

2015-01-06 Thread Gregory Farnum
So last month a bunch of librados functions around watch-notify were marked as deprecated, and because RBD still uses them everything went yellow on the gitbuilders. I believe we're expecting a patch series to move to the new APIs pretty soon, but was wondering when. In particular, a spot check

Re: replace fiemap with seek_data/seek_hole?

2015-01-06 Thread Gregory Farnum
On Tue, Jan 6, 2015 at 8:44 AM, Sage Weil sw...@redhat.com wrote: Hey, In an exchange on linux-fsdevel yesterday it became clear that even when FIEMAP isn't buggy it's not a good interface to build a map of sparse files. For example, XFS defrag or other future fs features may muck with

Re: Ceph backports

2015-01-05 Thread Gregory Farnum
On Mon, Jan 5, 2015 at 4:16 AM, Loic Dachary l...@dachary.org wrote: On 05/01/2015 13:03, John Spray wrote: Sounds sane -- is the new plan to always do backports via this process? i.e. if I see a backport PR which has not been through integration testing, should I refrain from merging it?

Re: Ceph backports

2015-01-05 Thread Gregory Farnum
On Mon, Jan 5, 2015 at 4:12 PM, Loic Dachary l...@dachary.org wrote: :-) This process is helpful if it allows me to help a little more than I currently do with the backport process. It would be a loss if the end result is that everyone cares less about backports. My primary incentive for

Re: CephFS hangs when writing 10GB files in loop

2014-12-17 Thread Gregory Farnum
On Wed, Dec 17, 2014 at 8:35 AM, Wido den Hollander w...@42on.com wrote: Hi, Today I've been playing with CephFS and the morning started great with CephFS playing along just fine. Some information first: - Ceph 0.89 - Linux kernel 3.18 - Ceph fuse 0.89 - One Active MDS, one Standby

Re: Accessing file layout information

2014-12-15 Thread Gregory Farnum
On Mon, Dec 15, 2014 at 10:54 AM, Atchley, Scott atchle...@ornl.gov wrote: Hi all, For a given file in cephfs, I would like to determine: 1) the number of PGs 2) the PG IDs 3) the offsets handles by each PG 4) the stripe unit (i.e. bytes per block of data) preferably using a C API. I

Re: Accessing file layout information

2014-12-15 Thread Gregory Farnum
On Mon, Dec 15, 2014 at 12:06 PM, Atchley, Scott atchle...@ornl.gov wrote: On Dec 15, 2014, at 2:42 PM, Gregory Farnum g...@gregs42.com wrote: On Mon, Dec 15, 2014 at 10:54 AM, Atchley, Scott atchle...@ornl.gov wrote: Hi all, For a given file in cephfs, I would like to determine: 1

Re: Accessing file layout information

2014-12-15 Thread Gregory Farnum
On Mon, Dec 15, 2014 at 1:29 PM, Atchley, Scott atchle...@ornl.gov wrote: On Dec 15, 2014, at 4:10 PM, Gregory Farnum g...@gregs42.com wrote: On Mon, Dec 15, 2014 at 12:06 PM, Atchley, Scott atchle...@ornl.gov wrote: On Dec 15, 2014, at 2:42 PM, Gregory Farnum g...@gregs42.com wrote: On Mon

Re: performance latency analysis with LTTng

2014-12-11 Thread Gregory Farnum
On Wed, Dec 10, 2014 at 10:54 PM, Blinick, Stephen L stephen.l.blin...@intel.com wrote: Hi Andreas, Thanks for providing this info last week. I've reviewed the data, the annotated spreadsheet is very helpful to interpret the results. I do think we're seeing similar measurements even

Re: Is there any documents to describe the architecture of ceph unit test based on gtest

2014-12-09 Thread Gregory Farnum
On Tue, Dec 9, 2014 at 1:50 AM, Nicheal zay11...@gmail.com wrote: Hi, all Is any guideline that describes how to run the ceph unit test, and its basic architecture? You can run them all by executing make check [-j N]. The executables run as part of that are specified in the makefiles

Re: Lock Constrains about Fast Dispatch for Messenger

2014-12-03 Thread Gregory Farnum
On Tue, Dec 2, 2014 at 10:39 PM, Haomai Wang haomaiw...@gmail.com wrote: Another question: we have lossy, server, standy and resetcheck as policy for Messenger, but in practice there exists some overlap semantics among these fields. For example, server and standy plays the same function in

Re: Lock Constrains about Fast Dispatch for Messenger

2014-12-02 Thread Gregory Farnum
On Tue, Dec 2, 2014 at 12:26 AM, Haomai Wang haomaiw...@gmail.com wrote: Hi Gregoy and Sage, I'm just writing Messenger's unit tests to ensure that SimpleMessenger and AsyncMessenger behavior same and expected. I think the most unclear thing is the lock rule of fast dispatch. When

Re: LTTng tracing: ReplicatedPG::log_operation

2014-12-02 Thread Gregory Farnum
On Tue, Dec 2, 2014 at 10:17 AM, Andreas Bluemle andreas.blue...@itxperts.de wrote: Hi, during code profiling using LTTng, I encounter that during processing of write requests to the cluster, the ceph-osd spends a lot of time in the ReplicatedPG::log_operation before the the actual writes to

Re: Ceph reviewers team

2014-11-25 Thread Gregory Farnum
On Tue, Nov 25, 2014 at 2:32 AM, Loic Dachary l...@dachary.org wrote: [ceph-devel cc'ed] Hi Sam, I created a Reviewer team for https://github.com/ceph/ceph/ : the members do not have special rights to the repository, but they can be assigned pull request which is convenient for

Re: Ceph Cinder HA

2014-11-10 Thread Gregory Farnum
On Mon, Nov 10, 2014 at 3:50 AM, Giulio Fidente gfide...@redhat.com wrote: On 11/10/2014 11:51 AM, Loic Dachary wrote: On 10/11/2014 11:34, Sebastien Han wrote: Indeed, that’s what we discussed with Giulio, using HAProxy for the Monitors is a bit overkill since the client is natively able

Re: Pull requests : speed up the reviews

2014-11-10 Thread Gregory Farnum
On Mon, Nov 10, 2014 at 11:50 AM, Loic Dachary l...@dachary.org wrote: Hi Greg, On Mon, 10 Nov 2014, Gregory Farnum wrote: I suppose one way of handling this might be to ask everybody to dedicate a small amount of time to reviews (as you suggest), but to emphasize PR management as much

Re: Debug messages of osd_ping and ping_reply

2014-10-30 Thread Gregory Farnum
On Thu, Oct 30, 2014 at 1:26 AM, Varada Kari varada.k...@sandisk.com wrote: Hi all, We have a test cluster with ~500 OSD's. This cluster has multiple rbd images mapped to multiple client machines. But from past two days, only one rbd image was serving the IO's rest of the images were not

Re: Any way to list the keysvalues of omap in certain range?

2014-10-27 Thread Gregory Farnum
Software Engineer #42 @ http://inktank.com | http://ceph.com On Mon, Oct 27, 2014 at 1:52 AM, Plato Zhang sango...@gmail.com wrote: Hi, all! We are trying to utilize omap to store kv map. However, we couldn't find a proper way to scan a certain key range of omap. We do can use arg

Re: Teuthology smoke test case(tasks/rados_python.yaml) failed on giant

2014-10-20 Thread Gregory Farnum
It sounds like you're running the master branch ceph-qa-suite tests against the giant branch of Ceph. The tests should pass if you resolve that. If not, or if you have some particular need for this function in the Giant release, you (or someone you work with) can submit a backport Pull Request,

Re: snap_trimming + backfilling is inefficient with many purged_snaps

2014-10-16 Thread Gregory Farnum
On Thu, Oct 16, 2014 at 2:04 AM, Florian Haas flor...@hastexo.com wrote: Hi Greg, sorry, this somehow got stuck in my drafts folder. On Tue, Sep 23, 2014 at 10:00 PM, Gregory Farnum g...@inktank.com wrote: On Tue, Sep 23, 2014 at 6:20 AM, Florian Haas flor...@hastexo.com wrote: On Mon, Sep

Re: Firefly maintenance release schedule

2014-10-15 Thread Gregory Farnum
On Wed, Oct 15, 2014 at 9:39 AM, Dmitry Borodaenko dborodae...@mirantis.com wrote: On Tue, Sep 30, 2014 at 6:49 PM, Dmitry Borodaenko dborodae...@mirantis.com wrote: Last stable Firefly release (v0.80.5) was tagged on July 29 (over 2 months ago). Since then, there were twice as many commits

Re: WriteBack Throttle kill the performace of the disk

2014-10-13 Thread Gregory Farnum
On Mon, Oct 13, 2014 at 6:29 AM, Mark Nelson mark.nel...@inktank.com wrote: On 10/13/2014 05:18 AM, Nicheal wrote: Hi, I'm currently finding that enable WritebackThrottle lead to lower IOPS for large number of small io. Since WritebackThrottle calls fdatasync(fd) to flush an object content

Re: why index (collectionIndex) need a lock?

2014-10-10 Thread Gregory Farnum
we prevent the Collection (merge or split) when modifying omap keys? Actually, the information in omap is pg_log, pg_info, and sometimes, object XATTR, right? 2014-10-01 1:23 GMT+08:00 Gregory Farnum g...@inktank.com: On Tue, Sep 30, 2014 at 12:08 AM, Nicheal zay11...@gmail.com wrote: Dear

Re: [Ceph-qa] New Defects reported by Coverity Scan for ceph

2014-10-09 Thread Gregory Farnum
On Thu, Oct 9, 2014 at 6:23 AM, scan-ad...@coverity.com wrote: Hi, Please find the latest report on new defect(s) introduced to ceph found with Coverity Scan. 3 new defect(s) introduced to ceph found with Coverity Scan. 4 defect(s), reported by Coverity Scan earlier, were marked fixed in

Re: why index (collectionIndex) need a lock?

2014-09-30 Thread Gregory Farnum
On Tue, Sep 30, 2014 at 12:08 AM, Nicheal zay11...@gmail.com wrote: Dear develops, I go through the files (hashIndex, LFNIndex and CollectionIndex), I cannot find any parameters need to grasp a lock. And basically, (hashIndex, LFNIndex and CollectionIndex) is used to manage a collection of

Re: New Defects reported by Coverity Scan for ceph (fwd)

2014-09-30 Thread Gregory Farnum
On Tue, Sep 30, 2014 at 6:59 AM, Sage Weil sw...@redhat.com wrote: Looks like recent changes from Greg, Loic, and I. -- Forwarded message -- From: scan-ad...@coverity.com To: undisclosed-recipients:; Cc: Date: Tue, 30 Sep 2014 06:21:08 -0700 Subject: New Defects reported by

Re: snap_trimming + backfilling is inefficient with many purged_snaps

2014-09-23 Thread Gregory Farnum
On Tue, Sep 23, 2014 at 6:20 AM, Florian Haas flor...@hastexo.com wrote: On Mon, Sep 22, 2014 at 7:06 PM, Florian Haas flor...@hastexo.com wrote: On Sun, Sep 21, 2014 at 9:52 PM, Sage Weil sw...@redhat.com wrote: On Sun, 21 Sep 2014, Florian Haas wrote: So yes, I think your patch absolutely

Re: set_alloc_hint old osds

2014-09-11 Thread Gregory Farnum
On Thu, Sep 11, 2014 at 1:19 PM, Samuel Just sam.j...@inktank.com wrote: http://tracker.ceph.com/issues/9419 librbd unconditionally sends set_alloc_hint. Do we require that users upgrade the osds first? Also, should the primary respond with ENOTSUPP if any replicas don't support it?

Re: set_alloc_hint old osds

2014-09-11 Thread Gregory Farnum
have any actual affect on the contents of the object. -Sam On Thu, Sep 11, 2014 at 1:30 PM, Gregory Farnum g...@inktank.com wrote: On Thu, Sep 11, 2014 at 1:19 PM, Samuel Just sam.j...@inktank.com wrote: http://tracker.ceph.com/issues/9419 librbd unconditionally sends set_alloc_hint. Do we

Re: set_alloc_hint old osds

2014-09-11 Thread Gregory Farnum
On Thu, Sep 11, 2014 at 1:40 PM, Gregory Farnum g...@inktank.com wrote: Does the hint not go into the pg log? Which could be retried on an older OSD? On Thu, Sep 11, 2014 at 1:33 PM, Samuel Just sam.j...@inktank.com wrote: That part is harmless, the transaction would be recreated for the new

Re: Question to RWLock reverse DNS ip=hostname

2014-09-09 Thread Gregory Farnum
On Tue, Sep 9, 2014 at 10:50 AM, Sage Weil sw...@redhat.com wrote: On Tue, 9 Sep 2014, Andreas Joachim Peters wrote: Hi, by chance I had a look to the RWLock class. To my best knowledge the way you create RW locks it defaults to writer-starvation e.g. all readers will always jump a head of a

Re: ceph data locality

2014-09-08 Thread Gregory Farnum
On Thu, Sep 4, 2014 at 12:16 AM, Johnu George (johnugeo) johnu...@cisco.com wrote: Hi All, I was reading more on Hadoop over ceph. I heard from Noah that tuning of Hadoop on Ceph is going on. I am just curious to know if there is any reason to keep default object size as 64MB. Is it

Re: ceph data locality

2014-09-08 Thread Gregory Farnum
On Mon, Sep 8, 2014 at 3:53 PM, Johnu George (johnugeo) johnu...@cisco.com wrote: Hi Greg, Thanks. Can you explain more on Ceph *does* export locations so the follow-up jobs can be scheduled appropriately”? Thanks, Johnu On 9/8/14, 12:51 PM, Gregory Farnum g...@inktank.com wrote: On Thu

Re: new developer - where to start contributing

2014-08-29 Thread Gregory Farnum
On Thu, Aug 28, 2014 at 7:12 AM, Kostiantyn Danylov kdani...@mirantis.com wrote: Hi all, My name is Kostiantyn Danylov, I'm C++/C/python developer with ~9 year of experience, also I'm interesting in distributed systems. I would like to contribute to ceph. I already read documents at

Re: OSDMap checksums

2014-08-22 Thread Gregory Farnum
On Thu, Aug 21, 2014 at 5:38 PM, Sage Weil sw...@redhat.com wrote: On Tue, 19 Aug 2014, Gregory Farnum wrote: As far as I can tell, checksumming incrementals are good for two things besides detecting bit flips: 1) It's easy to extend to signing the Incremental, which is more secure 2

Re: watch/notify changes

2014-08-22 Thread Gregory Farnum
On Thu, Aug 21, 2014 at 3:34 PM, Sage Weil sw...@redhat.com wrote: Sam and Josh and I discussed the state of watch/notify a couple weeks back. Here are our notes: http://pad.ceph.com/p/watch-notify I've mapped most of these to tickets or bugs and noted them in the pad. Ignore the

Re: watch/notify changes

2014-08-22 Thread Gregory Farnum
On Fri, Aug 22, 2014 at 11:22 AM, Sage Weil sw...@redhat.com wrote: On Fri, 22 Aug 2014, Gregory Farnum wrote: On Thu, Aug 21, 2014 at 3:34 PM, Sage Weil sw...@redhat.com wrote: Sam and Josh and I discussed the state of watch/notify a couple weeks back. Here are our notes: http

Re: watch/notify changes

2014-08-22 Thread Gregory Farnum
On Fri, Aug 22, 2014 at 2:30 PM, Sage Weil sw...@redhat.com wrote: On Fri, 22 Aug 2014, Gregory Farnum wrote: Whereas if the notify timeout is the same time length as a watch timeout, we can affirmatively know on a notify reply (with or without error return codes) that every client has either

Re: watch/notify changes

2014-08-22 Thread Gregory Farnum
On Fri, Aug 22, 2014 at 3:59 PM, Sage Weil sw...@redhat.com wrote: On Fri, 22 Aug 2014, Gregory Farnum wrote: Can you give some examples of situations in which an eventual delivery is a useful building block for something other than I know this was delivered? I'm having trouble coming up

Re: Ceph authentication/authorization paradignms

2014-08-21 Thread Gregory Farnum
On Wed, Aug 20, 2014 at 3:20 PM, Christopher R. Hertel c...@redhat.com wrote: [At end...] - Original Message - From: Gregory Farnum g...@inktank.com To: Christopher R. Hertel c...@redhat.com Cc: ceph-devel@vger.kernel.org Sent: Tuesday, August 19, 2014 4:57:59 PM Subject: Re: Ceph

Re: [ceph-users] wired output of ceph df : Firefly 0.80.5

2014-08-21 Thread Gregory Farnum
Yeah, this is pretty weird. Can you run df on each of your OSDs and see if any of them individually have a strange output? -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Tue, Aug 12, 2014 at 6:52 AM, Karan Singh karan.si...@csc.fi wrote: Forgot to mention i am observing

Re: mon dependency on OSD MonClient using OSD's client_messenger

2014-08-21 Thread Gregory Farnum
On Wed, Aug 6, 2014 at 9:40 AM, Matt W. Benjamin m...@cohortfs.com wrote: Hi, We noticed that OSDs use their client_messenger, rather than their potentially dedicated cluster_messenger, to drive MonClient. Then we found that the Mon's setup of the OSDMap depends on this. Can someone

Re: Wireshark Dissector and the Future

2014-08-21 Thread Gregory Farnum
On Sun, Aug 10, 2014 at 5:22 AM, Kevin Cox kevin...@kevincox.ca wrote: -BEGIN PGP SIGNED MESSAGE- Hash: SHA256 Hello, This summer as a GSOC project I have implemented a Wireshark dissector for Ceph. It has been committed to Wireshark master and will be available in the next

Re: Building a tool which links with librados

2014-08-21 Thread Gregory Farnum
On Thu, Aug 21, 2014 at 4:37 PM, David Zafman david.zaf...@inktank.com wrote: Has anyone seen anything like this from an application linked with librados using valgrind? Or a Segmentation fault on exit from such an application? Invalid free() / delete / delete[] / realloc() at

Re: size and depth of Tree Bucket

2014-08-19 Thread Gregory Farnum
On Wed, Aug 13, 2014 at 4:09 AM, Viral Mehta viral@gmail.com wrote: Hi, I was looking at ceph/crush/builder.c crush_make_tree_bucket() function, I would like to understand what size argument refers to ? Generally speaking Tree Size is Total Number of the nodes. And this would be the same

Re: Ceph authentication/authorization paradignms

2014-08-19 Thread Gregory Farnum
On Thu, Aug 14, 2014 at 10:10 AM, Christopher R. Hertel c...@redhat.com wrote: Ceph Folks: Hi. I am new to Ceph but I've been around in the Open Source world for a while, working on a variety of different projects (mostly Windows Interop stuff). I was asked to take a look at the

Re: Simpler LRC configuration

2014-08-19 Thread Gregory Farnum
On Fri, Aug 15, 2014 at 9:44 AM, Loic Dachary l...@dachary.org wrote: Hi, Based on Andreas suggestions for Xorbas configuration (in a previous thread) and the last G/H discussions ( https://wiki.ceph.com/Planning/CDS/CDS_Giant_and_Hammer_%28Jun_2014%29 ), here is a simpler configuration

Re: OSD suicide after being down/in for one day as it needs to search large amount of objects

2014-08-19 Thread Gregory Farnum
On Mon, Aug 18, 2014 at 11:30 PM, Guang Yang yguan...@outlook.com wrote: Hi ceph-devel, David (cc’ed) reported a bug (http://tracker.ceph.com/issues/9128) which we came across in our test cluster during our failure testing, basically the way to reproduce it was to leave one OSD daemon down

Re: OSDMap checksums

2014-08-19 Thread Gregory Farnum
On Tue, Aug 19, 2014 at 3:43 PM, Sage Weil sw...@redhat.com wrote: We have had a range of bugs come up in the past because OSDs or mons have been running different versions of the code and have encoded different variations of the same OSDMap epoch. When two nodes in the system disagree about

Re: OSDMap checksums

2014-08-19 Thread Gregory Farnum
On Tue, Aug 19, 2014 at 5:32 PM, Sage Weil sw...@redhat.com wrote: On Tue, 19 Aug 2014, Gregory Farnum wrote: On Tue, Aug 19, 2014 at 3:43 PM, Sage Weil sw...@redhat.com wrote: We have had a range of bugs come up in the past because OSDs or mons have been running different versions

Re: OSDMap checksums

2014-08-19 Thread Gregory Farnum
On Tue, Aug 19, 2014 at 9:49 PM, Sage Weil sw...@redhat.com wrote: Right, so let's talk about how we get into that situation: 1) Our existing OSDMap is bad. a) We were never correct b) ...we went bad and didn't notice? 2) The Incremental we got is bad. a) It's not the original

Re: KeyFileStore ?

2014-07-31 Thread Gregory Farnum
On Thu, Jul 31, 2014 at 1:25 AM, Sage Weil sw...@redhat.com wrote: After the latest set of bug fixes to the FileStore file naming code I am newly inspired to replace it with something less complex. Right now I'm mostly thinking about HDDs, although some of this may map well onto hybrid

Re: [ceph-users] question about ApplyManager, SubmitManager and FileJournal

2014-07-31 Thread Gregory Farnum
These sorts of questions are good for ceph-devel@vger.kernel.org, which I've added. :) On Thu, Jul 31, 2014 at 12:24 PM, yuelongguang fasts...@163.com wrote: hi,all recently i dive into the source code, i am a little confused about them, maybe because of many threads,wait,seq. 1. what does

Re: giant and hammer dates

2014-07-30 Thread Gregory Farnum
On Tue, Jul 29, 2014 at 7:11 PM, Sage Weil sw...@redhat.com wrote: We've talked a bit about moving to a ~4 month (instead of 3 month) cadence. I'm still inclined in this direction because it means fewer stable releases that we will be maintaining and a longer and (hopefully) more productive

Re: [PATCH] ceph: use the OSD time as the object mtime instead of the client time

2014-07-29 Thread Gregory Farnum
On Tue, Jul 29, 2014 at 11:46 AM, Sage Weil sw...@redhat.com wrote: On Tue, 29 Jul 2014, Wang, Zhiqiang wrote: This fixes a bug when the time of the OSDs and clients are not synchronized (especially when client is ahead of OSD), and the cache tier dirty ratio reaches the threshold, the agent

Re: Disabling CRUSH for erasure code and doing custom placement

2014-07-23 Thread Gregory Farnum
On Tue, Jul 22, 2014 at 2:48 PM, Shayan Saeed shayansaee...@gmail.com wrote: Another question along the same lines. For erasure code, same as replicated files, the request goes through the primary member. Isn't it possible to send the request to any of the members and get the file. While this

Re: Adding a delay when restarting all OSDs on a host

2014-07-22 Thread Gregory Farnum
On Tue, Jul 22, 2014 at 6:19 AM, Wido den Hollander w...@42on.com wrote: Hi, Currently on Ubuntu with Upstart when you invoke a restart like this: $ sudo restart ceph-osd-all It will restart all OSDs at once, which can increase the load on the system a quite a bit. It's better to restart

Re: [RFC][PATCH] osd: Add local_connection to fast_dispatch in func _send_boot.

2014-07-21 Thread Gregory Farnum
On Sun, Jul 20, 2014 at 11:33 PM, Ma, Jianpeng jianpeng...@intel.com wrote: Hrm, I'd really like to see the startup sequence. I see the crash occurring, but I don't understand how it's happening — we test this pretty extensively so there must be something about your testing configuration

Re: Forcing Ceph into mapping all objects to a single PG

2014-07-21 Thread Gregory Farnum
On Mon, Jul 21, 2014 at 3:27 PM, Daniel Hofmann dan...@trvx.org wrote: Preamble: you might want to read the decent formatted version of this mail at: https://gist.github.com/daniel-j-h/2daae2237bb21596c97d Motivation -- Recently I was wondering how Ceph (i.e. Rados in particular)

Re: Disabling CRUSH for erasure code and doing custom placement

2014-07-18 Thread Gregory Farnum
On Fri, Jul 18, 2014 at 2:12 AM, Kaifeng Yao kaif...@yahoo-inc.com wrote: I was thinking about 'PG preferred' to allow binding a PG's placement to arbitrary OSDs. My angle is to make the PG more evenly distributed across OSDs, thus to potentially save ~20% cost. I am searching the 'pg

Re: [RFC][PATCH] osd: Add local_connection to fast_dispatch in func _send_boot.

2014-07-18 Thread Gregory Farnum
wrote: Hi Greg, The attachment is the log. Thanks! -Original Message- From: Gregory Farnum [mailto:g...@inktank.com] Sent: Thursday, July 17, 2014 3:41 AM To: Ma, Jianpeng Cc: ceph-devel@vger.kernel.org Subject: Re: [RFC][PATCH] osd: Add local_connection to fast_dispatch

Re: [ceph-users] v0.80.4 Firefly released

2014-07-16 Thread Gregory Farnum
On Wed, Jul 16, 2014 at 1:50 AM, James Harper ja...@ejbdigital.com.au wrote: Can you offer some comments on what the impact is likely to be to the data in an affected cluster? Should all data now be treated with suspicion and restored back to before the firefly upgrade? I am under the

Re: v0.80.4 Firefly released

2014-07-16 Thread Gregory Farnum
On Wed, Jul 16, 2014 at 2:22 AM, Christoph Hellwig h...@infradead.org wrote: On Tue, Jul 15, 2014 at 04:45:59PM -0700, Sage Weil wrote: This Firefly point release fixes an potential data corruption problem when ceph-osd daemons run on top of XFS and service Firefly librbd clients. A recently

Re: Read from clones

2014-07-16 Thread Gregory Farnum
FYI, this sounds like an issue the userspace client (and possibly the actual rbd class?) had as well: it looked at the HEAD parent_overlap field even when reading from snapshots. (I don't remember if parent_overlap is the actual parameter name, but you get the idea.) -Greg Software Engineer #42 @

<    1   2   3   4   5   6   7   8   9   10   >