On Wed, Feb 11, 2015 at 10:19 AM, Loic Dachary l...@dachary.org wrote:
On 11/02/2015 18:27, Gregory Farnum wrote:
Mmm. I'm happy to look at suites that get run this way but I'm
unlikely to notice them go by on the list if I'm not poked about them
— I generally filter out anything that has
On Tue, Feb 10, 2015 at 9:22 AM, Loic Dachary l...@dachary.org wrote:
On 10/02/2015 18:19, Yuri Weinstein wrote:
Loic,
The only difference between options if we run suits on merged dumpling vs
dumpling-backports first - is time.
We will have to run suites on the final branch after the
On Tue, Feb 10, 2015 at 10:04 AM, Loic Dachary l...@dachary.org wrote:
On 10/02/2015 18:29, Yuri Weinstein wrote:
On 10/02/2015 18:19, Yuri Weinstein wrote:
Loic,
The only difference between options if we run suits on merged dumpling vs
dumpling-backports first - is time.
We will have to
On Tue, Feb 10, 2015 at 10:26 AM, Sage Weil sw...@redhat.com wrote:
On Tue, 10 Feb 2015, Somnath Roy wrote:
Thanks Sam !
So, is it safe to do ordering if in a transaction *no*
remove/truncate/create/add call ?
For example, do we need to preserve ordering in case of the below
transaction ?
On Tue, Feb 10, 2015 at 10:33 AM, Loic Dachary l...@dachary.org wrote:
On 10/02/2015 19:25, Gregory Farnum wrote:
Now, as it happens there are some reasons to maintain a dumpling
branch that isn't part of backports. We've been doing a lot of work
lately to make the nightlies behave well
On Tue, Feb 10, 2015 at 10:55 AM, Stefan Priebe s.pri...@profihost.ag wrote:
Hello,
last year in june i already reported this but there was no real result.
(http://lists.ceph.com/pipermail/ceph-users-ceph.com/2014-July/041070.html)
I then had the hope that this will be fixed itself when
Haomai,
On Sun, Feb 8, 2015 at 2:22 AM, ceph.git ceph-com...@ceph.com wrote:
This is an automated email from the git hooks/post-receive script. It was
generated because a ref change was pushed to the repository containing
the project .
The branch, master has been updated
via
On Mon, Feb 9, 2015 at 7:33 AM, Yehuda Sadeh-Weinraub yeh...@redhat.com wrote:
- Original Message -
From: Sage Weil s...@newdream.net
To: Yehuda Sadeh-Weinraub yeh...@redhat.com
Cc: Ceph Development ceph-devel@vger.kernel.org
Sent: Monday, February 9, 2015 3:42:40 AM
Subject: Re:
On Sun, Feb 8, 2015 at 1:38 PM, Sage Weil sw...@redhat.com wrote:
Simon Leinen at Switch did a greaet post recently about the impact of
scrub on their cluster(s):
http://blog.simon.leinen.ch/2015/02/ceph-deep-scrubbing-impact.html
Basically the 2 week deep scrub interval kicks in on
Right.
So, memory usage of an OSD is usually linear in the number of PGs it
hosts. However, that memory can also grow based on at least one other
thing: the number of OSD Maps required to go through peering. It
*looks* to me like this is what you're running in to, not growth on
the number of
- Original Message -
From: Haomai Wang haomaiw...@gmail.com
To: Gregory Farnum gfar...@redhat.com
Cc: Sage Weil sw...@redhat.com, ceph-devel@vger.kernel.org
Sent: Friday, February 6, 2015 8:16:42 AM
Subject: Re: About in_seq, out_seq in Messenger
On Fri, Feb 6, 2015 at 10:47 PM
On Fri, Feb 6, 2015 at 1:46 PM, Yehuda Sadeh-Weinraub yeh...@redhat.com wrote:
I have been recently looking at implementing object expiration in rgw. First,
a brief description of the feature:
S3 provides mechanisms to expire objects, and/or to transition them into
different storage class.
On Tue, Feb 3, 2015 at 4:12 AM, Ding Dinghua dingdinghu...@gmail.com wrote:
Hi all:
I don't understand why SnapMapper::get_prefix static_cast snap
to unsigned:
string SnapMapper::get_prefix(snapid_t snap)
{
char buf[100];
int len = snprintf(
buf, sizeof(buf),
%.*X_,
The guys who would normally handle this have been traveling lately, so
they're probably behind on such things.
That said, Github pull requests are probably a more reliable
transmission channel than the mailing list. :)
-Greg
On Tue, Feb 3, 2015 at 11:31 AM, Yazen Ghannam yazen.ghan...@linaro.org
On Mon, Feb 2, 2015 at 7:00 AM, Loic Dachary l...@dachary.org wrote:
On 02/02/2015 14:48, Yan, Zheng wrote:
On Mon, Feb 2, 2015 at 9:18 PM, Loic Dachary l...@dachary.org wrote:
Hi,
http://pulpito.ceph.com/loic-2015-01-29_15:39:38-rbd-dumpling-backports---basic-multi/730029/
hangs on
On Wed, Jan 28, 2015 at 10:06 AM, Sage Weil s...@newdream.net wrote:
On Wed, 28 Jan 2015, John Spray wrote:
On Wed, Jan 28, 2015 at 5:23 PM, Gregory Farnum g...@gregs42.com wrote:
My concern is whether we as the FS are responsible for doing anything
more than storing and returning
On Wed, Jan 28, 2015 at 5:24 AM, John Spray john.sp...@redhat.com wrote:
We don't implement the GETFLAGS and SETFLAGS ioctls used for +i.
Adding the ioctls is pretty easy, but then we need somewhere to put
the flags. Currently we don't store a flags attribute on inodes,
but maybe we could
Nifty; it's good to have that sort of blog-style documentation about
the interface. Are you planning to do some work with it that you can
show off as well? :)
-Greg
On Tue, Jan 27, 2015 at 12:48 PM, Marcel Lauhoff m...@irq0.org wrote:
Hi,
I wrote an article about the object store API - How it
On Fri, Jan 23, 2015 at 1:43 PM, Sage Weil sw...@redhat.com wrote:
Background:
1) Way back when we made a task that would thrash the cache modes by
adding and removing the cache tier while ceph_test_rados was running.
This mostly worked, but would occasionally fail because we would
-
On Fri, Jan 23, 2015 at 2:18 PM, Sage Weil sw...@redhat.com wrote:
On Fri, 23 Jan 2015, Gregory Farnum wrote:
On Fri, Jan 23, 2015 at 1:43 PM, Sage Weil sw...@redhat.com wrote:
Background:
1) Way back when we made a task that would thrash the cache modes by
adding and removing the cache
On Sun, Jan 18, 2015 at 11:02 PM, Mykola Golub mgo...@mirantis.com wrote:
On Sun, Jan 18, 2015 at 10:33:05AM -0800, Sage Weil wrote:
On Sun, 18 Jan 2015, Mykola Golub wrote:
Hi Ceph,
Right now, for not a monitor leader, if a received command is not
supported locally, but is supported by
Sage, are these uncaught assertion errors something we normally
ignore? I'm not familiar with any code that tries to catch errors in
our standard init patterns, which is what looks to be the problem on
these new coverity issues in cephfs-table-tool.
-Greg
On Fri, Jan 16, 2015 at 6:39 AM,
On Fri, Jan 16, 2015 at 10:34 AM, Michael Sevilla
mikesevil...@gmail.com wrote:
On Thu, Jan 15, 2015 at 10:37 PM, Gregory Farnum g...@gregs42.com wrote:
On Thu, Jan 15, 2015 at 2:44 PM, Michael Sevilla mikesevil...@gmail.com
wrote:
Let me know if this works and/or you need anything else
On Fri, Jan 16, 2015 at 11:28 AM, Lipeng Wan lipengwa...@gmail.com wrote:
Dear all,
Does Ceph provide a way to collect object-level I/O access traces?
Specifically, can we collect the traces to record how many times each
object has been accessed (read, write, etc.) during a fixed period of
, ceph-mds.ceph-node1.log, etc., which
log file should I look at? Maybe the ceph-mds.ceph-node1.log?
Specifically, is there any keyword I can search in the log file to
locate the object operations?
Thanks!
LW
On Fri, Jan 16, 2015 at 4:21 PM, Gregory Farnum g...@gregs42.com wrote:
On Fri, Jan
?
Cheng
Cheng
On Thu, Jan 15, 2015 at 12:59 PM, Gregory Farnum g...@gregs42.com wrote:
On Thu, Jan 15, 2015 at 9:53 AM, Cheng Cheng ccheng@gmail.com wrote:
Hi Ceph,
I am wondering is there a mechanism to prioritize the
rbd_aio_write/rbd_aio_read I/Os? Currently all RBD I/Os
On Thu, Jan 15, 2015 at 9:53 AM, Cheng Cheng ccheng@gmail.com wrote:
Hi Ceph,
I am wondering is there a mechanism to prioritize the
rbd_aio_write/rbd_aio_read I/Os? Currently all RBD I/Os are issued in FIFO to
rados layer, and there is NO QoS mechanism to control the priority of these
On Thu, Jan 15, 2015 at 9:44 AM, Sage Weil sw...@redhat.com wrote:
In addition (or instead of) making the API harder to fat-finger, we could
also add a mon config option like
mon allow pool deletion = false
that defaults off. Then, to delete any pool, you need to update ceph.conf
and
Oh, I think it might be fine to require setting a config option before
you delete stuff; I just don't want to prevent the option from being
injectable. :)
On Thu, Jan 15, 2015 at 10:07 AM, Sage Weil sw...@redhat.com wrote:
On Thu, 15 Jan 2015, Gregory Farnum wrote:
On Thu, Jan 15, 2015 at 9:44
On Thu, Jan 15, 2015 at 2:44 PM, Michael Sevilla mikesevil...@gmail.com wrote:
Let me know if this works and/or you need anything else:
https://www.dropbox.com/s/fq47w6jebnyluu0/lookup-logs.tar.gz?dl=0
Beware - the clients were on debug=10. Also, I tried this with the
kernel client and it is
Can you post the full logs somewhere to look at? These bits aren't
very helpful on their own (except to say, yes, the client cleared its
I_COMPLETE for some reason).
On Tue, Jan 13, 2015 at 3:45 PM, Michael Sevilla mikesevil...@gmail.com wrote:
On Tue, Jan 13, 2015 at 11:13 AM, Gregory Farnum g
On Mon, Jan 12, 2015 at 10:17 PM, Michael Sevilla
mikesevil...@gmail.com wrote:
I can't get consistent performance with 1 MDS. I have 2 clients create
100,000 files (separate directories) in a CephFS mount. I ran the
experiment 5 times (deleting the pools/fs and restarting the MDS in
between
On Fri, Jan 9, 2015 at 7:20 AM, Sage Weil sw...@redhat.com wrote:
Should we drop this entirely in hammer?
Yes!
If I remember correctly all of
the layout stuff is fully supported using virtual xattrs and standard
tools. The only thing left is the tool that shows you how file blocks map
to
On Fri, Jan 9, 2015 at 10:00 AM, Sage Weil sw...@redhat.com wrote:
On Fri, 9 Jan 2015, Gregory Farnum wrote:
On Fri, Jan 9, 2015 at 7:20 AM, Sage Weil sw...@redhat.com wrote:
Should we drop this entirely in hammer?
Yes!
If I remember correctly all of
the layout stuff is fully supported
On Fri, Jan 9, 2015 at 10:08 AM, Sage Weil sw...@redhat.com wrote:
On Fri, 9 Jan 2015, Gregory Farnum wrote:
On Fri, Jan 9, 2015 at 10:00 AM, Sage Weil sw...@redhat.com wrote:
On Fri, 9 Jan 2015, Gregory Farnum wrote:
On Fri, Jan 9, 2015 at 7:20 AM, Sage Weil sw...@redhat.com wrote
On Tue, Jan 6, 2015 at 12:39 AM, Loic Dachary l...@dachary.org wrote:
On 06/01/2015 01:22, Gregory Farnum wrote:
On Mon, Jan 5, 2015 at 4:12 PM, Loic Dachary l...@dachary.org wrote:
:-) This process is helpful if it allows me to help a little more than I
currently do with the backport
So last month a bunch of librados functions around watch-notify were
marked as deprecated, and because RBD still uses them everything went
yellow on the gitbuilders. I believe we're expecting a patch series to
move to the new APIs pretty soon, but was wondering when.
In particular, a spot check
On Tue, Jan 6, 2015 at 8:44 AM, Sage Weil sw...@redhat.com wrote:
Hey,
In an exchange on linux-fsdevel yesterday it became clear that even when
FIEMAP isn't buggy it's not a good interface to build a map of sparse
files. For example, XFS defrag or other future fs features may muck with
On Mon, Jan 5, 2015 at 4:16 AM, Loic Dachary l...@dachary.org wrote:
On 05/01/2015 13:03, John Spray wrote:
Sounds sane -- is the new plan to always do backports via this
process? i.e. if I see a backport PR which has not been through
integration testing, should I refrain from merging it?
On Mon, Jan 5, 2015 at 4:12 PM, Loic Dachary l...@dachary.org wrote:
:-) This process is helpful if it allows me to help a little more than I
currently do with the backport process. It would be a loss if the end result
is that everyone cares less about backports. My primary incentive for
On Wed, Dec 17, 2014 at 8:35 AM, Wido den Hollander w...@42on.com wrote:
Hi,
Today I've been playing with CephFS and the morning started great with
CephFS playing along just fine.
Some information first:
- Ceph 0.89
- Linux kernel 3.18
- Ceph fuse 0.89
- One Active MDS, one Standby
On Mon, Dec 15, 2014 at 10:54 AM, Atchley, Scott atchle...@ornl.gov wrote:
Hi all,
For a given file in cephfs, I would like to determine:
1) the number of PGs
2) the PG IDs
3) the offsets handles by each PG
4) the stripe unit (i.e. bytes per block of data)
preferably using a C API.
I
On Mon, Dec 15, 2014 at 12:06 PM, Atchley, Scott atchle...@ornl.gov wrote:
On Dec 15, 2014, at 2:42 PM, Gregory Farnum g...@gregs42.com wrote:
On Mon, Dec 15, 2014 at 10:54 AM, Atchley, Scott atchle...@ornl.gov wrote:
Hi all,
For a given file in cephfs, I would like to determine:
1
On Mon, Dec 15, 2014 at 1:29 PM, Atchley, Scott atchle...@ornl.gov wrote:
On Dec 15, 2014, at 4:10 PM, Gregory Farnum g...@gregs42.com wrote:
On Mon, Dec 15, 2014 at 12:06 PM, Atchley, Scott atchle...@ornl.gov wrote:
On Dec 15, 2014, at 2:42 PM, Gregory Farnum g...@gregs42.com wrote:
On Mon
On Wed, Dec 10, 2014 at 10:54 PM, Blinick, Stephen L
stephen.l.blin...@intel.com wrote:
Hi Andreas, Thanks for providing this info last week. I've reviewed the
data, the annotated spreadsheet is very helpful to interpret the results.
I do think we're seeing similar measurements even
On Tue, Dec 9, 2014 at 1:50 AM, Nicheal zay11...@gmail.com wrote:
Hi, all
Is any guideline that describes how to run the ceph unit test, and its
basic architecture?
You can run them all by executing make check [-j N]. The executables
run as part of that are specified in the makefiles
On Tue, Dec 2, 2014 at 10:39 PM, Haomai Wang haomaiw...@gmail.com wrote:
Another question:
we have lossy, server, standy and resetcheck as policy for
Messenger, but in practice there exists some overlap semantics among
these fields. For example, server and standy plays the same
function in
On Tue, Dec 2, 2014 at 12:26 AM, Haomai Wang haomaiw...@gmail.com wrote:
Hi Gregoy and Sage,
I'm just writing Messenger's unit tests to ensure that SimpleMessenger and
AsyncMessenger behavior same and expected.
I think the most unclear thing is the lock rule of fast dispatch. When
On Tue, Dec 2, 2014 at 10:17 AM, Andreas Bluemle
andreas.blue...@itxperts.de wrote:
Hi,
during code profiling using LTTng, I encounter that during
processing of write requests to the cluster, the ceph-osd
spends a lot of time in the ReplicatedPG::log_operation
before the the actual writes to
On Tue, Nov 25, 2014 at 2:32 AM, Loic Dachary l...@dachary.org wrote:
[ceph-devel cc'ed]
Hi Sam,
I created a Reviewer team for https://github.com/ceph/ceph/ : the members
do not have special rights to the repository, but they can be assigned pull
request which is convenient for
On Mon, Nov 10, 2014 at 3:50 AM, Giulio Fidente gfide...@redhat.com wrote:
On 11/10/2014 11:51 AM, Loic Dachary wrote:
On 10/11/2014 11:34, Sebastien Han wrote:
Indeed, that’s what we discussed with Giulio, using HAProxy for the
Monitors is a bit overkill since the client is natively able
On Mon, Nov 10, 2014 at 11:50 AM, Loic Dachary l...@dachary.org wrote:
Hi Greg,
On Mon, 10 Nov 2014, Gregory Farnum wrote:
I suppose one way of handling this might be to ask everybody to
dedicate a small amount of time to reviews (as you suggest), but to
emphasize PR management as much
On Thu, Oct 30, 2014 at 1:26 AM, Varada Kari varada.k...@sandisk.com wrote:
Hi all,
We have a test cluster with ~500 OSD's. This cluster has multiple rbd images
mapped to multiple client machines. But from past two days, only one rbd
image was serving the IO's rest of the images were not
Software Engineer #42 @ http://inktank.com | http://ceph.com
On Mon, Oct 27, 2014 at 1:52 AM, Plato Zhang sango...@gmail.com wrote:
Hi, all!
We are trying to utilize omap to store kv map.
However, we couldn't find a proper way to scan a certain key range of omap.
We do can use arg
It sounds like you're running the master branch ceph-qa-suite tests
against the giant branch of Ceph. The tests should pass if you resolve
that.
If not, or if you have some particular need for this function in the
Giant release, you (or someone you work with) can submit a backport
Pull Request,
On Thu, Oct 16, 2014 at 2:04 AM, Florian Haas flor...@hastexo.com wrote:
Hi Greg,
sorry, this somehow got stuck in my drafts folder.
On Tue, Sep 23, 2014 at 10:00 PM, Gregory Farnum g...@inktank.com wrote:
On Tue, Sep 23, 2014 at 6:20 AM, Florian Haas flor...@hastexo.com wrote:
On Mon, Sep
On Wed, Oct 15, 2014 at 9:39 AM, Dmitry Borodaenko
dborodae...@mirantis.com wrote:
On Tue, Sep 30, 2014 at 6:49 PM, Dmitry Borodaenko
dborodae...@mirantis.com wrote:
Last stable Firefly release (v0.80.5) was tagged on July 29 (over 2
months ago). Since then, there were twice as many commits
On Mon, Oct 13, 2014 at 6:29 AM, Mark Nelson mark.nel...@inktank.com wrote:
On 10/13/2014 05:18 AM, Nicheal wrote:
Hi,
I'm currently finding that enable WritebackThrottle lead to lower IOPS
for large number of small io. Since WritebackThrottle calls
fdatasync(fd) to flush an object content
we prevent the Collection (merge or split) when
modifying omap keys? Actually, the information in omap is pg_log,
pg_info, and sometimes, object XATTR, right?
2014-10-01 1:23 GMT+08:00 Gregory Farnum g...@inktank.com:
On Tue, Sep 30, 2014 at 12:08 AM, Nicheal zay11...@gmail.com wrote:
Dear
On Thu, Oct 9, 2014 at 6:23 AM, scan-ad...@coverity.com wrote:
Hi,
Please find the latest report on new defect(s) introduced to ceph found with
Coverity Scan.
3 new defect(s) introduced to ceph found with Coverity Scan.
4 defect(s), reported by Coverity Scan earlier, were marked fixed in
On Tue, Sep 30, 2014 at 12:08 AM, Nicheal zay11...@gmail.com wrote:
Dear develops,
I go through the files (hashIndex, LFNIndex and CollectionIndex), I
cannot find any parameters need to grasp a lock. And basically,
(hashIndex, LFNIndex and CollectionIndex) is used to manage a
collection of
On Tue, Sep 30, 2014 at 6:59 AM, Sage Weil sw...@redhat.com wrote:
Looks like recent changes from Greg, Loic, and I.
-- Forwarded message --
From: scan-ad...@coverity.com
To: undisclosed-recipients:;
Cc:
Date: Tue, 30 Sep 2014 06:21:08 -0700
Subject: New Defects reported by
On Tue, Sep 23, 2014 at 6:20 AM, Florian Haas flor...@hastexo.com wrote:
On Mon, Sep 22, 2014 at 7:06 PM, Florian Haas flor...@hastexo.com wrote:
On Sun, Sep 21, 2014 at 9:52 PM, Sage Weil sw...@redhat.com wrote:
On Sun, 21 Sep 2014, Florian Haas wrote:
So yes, I think your patch absolutely
On Thu, Sep 11, 2014 at 1:19 PM, Samuel Just sam.j...@inktank.com wrote:
http://tracker.ceph.com/issues/9419
librbd unconditionally sends set_alloc_hint. Do we require that users
upgrade the osds first? Also, should the primary respond with
ENOTSUPP if any replicas don't support it?
have any actual affect on the contents of the object.
-Sam
On Thu, Sep 11, 2014 at 1:30 PM, Gregory Farnum g...@inktank.com wrote:
On Thu, Sep 11, 2014 at 1:19 PM, Samuel Just sam.j...@inktank.com wrote:
http://tracker.ceph.com/issues/9419
librbd unconditionally sends set_alloc_hint. Do we
On Thu, Sep 11, 2014 at 1:40 PM, Gregory Farnum g...@inktank.com wrote:
Does the hint not go into the pg log? Which could be retried on an older OSD?
On Thu, Sep 11, 2014 at 1:33 PM, Samuel Just sam.j...@inktank.com wrote:
That part is harmless, the transaction would be recreated for the new
On Tue, Sep 9, 2014 at 10:50 AM, Sage Weil sw...@redhat.com wrote:
On Tue, 9 Sep 2014, Andreas Joachim Peters wrote:
Hi,
by chance I had a look to the RWLock class. To my best knowledge the way
you create RW locks it defaults to writer-starvation e.g. all readers
will always jump a head of a
On Thu, Sep 4, 2014 at 12:16 AM, Johnu George (johnugeo)
johnu...@cisco.com wrote:
Hi All,
I was reading more on Hadoop over ceph. I heard from Noah that
tuning of Hadoop on Ceph is going on. I am just curious to know if there
is any reason to keep default object size as 64MB. Is it
On Mon, Sep 8, 2014 at 3:53 PM, Johnu George (johnugeo)
johnu...@cisco.com wrote:
Hi Greg,
Thanks. Can you explain more on Ceph *does* export locations so
the follow-up jobs can be scheduled appropriately”?
Thanks,
Johnu
On 9/8/14, 12:51 PM, Gregory Farnum g...@inktank.com wrote:
On Thu
On Thu, Aug 28, 2014 at 7:12 AM, Kostiantyn Danylov
kdani...@mirantis.com wrote:
Hi all,
My name is Kostiantyn Danylov, I'm C++/C/python developer with ~9 year of
experience, also I'm interesting in distributed systems. I would like to
contribute to ceph. I already read documents at
On Thu, Aug 21, 2014 at 5:38 PM, Sage Weil sw...@redhat.com wrote:
On Tue, 19 Aug 2014, Gregory Farnum wrote:
As far as I can tell, checksumming incrementals are good for two
things besides detecting bit flips:
1) It's easy to extend to signing the Incremental, which is more secure
2
On Thu, Aug 21, 2014 at 3:34 PM, Sage Weil sw...@redhat.com wrote:
Sam and Josh and I discussed the state of watch/notify a couple weeks
back. Here are our notes:
http://pad.ceph.com/p/watch-notify
I've mapped most of these to tickets or bugs and noted them in the pad.
Ignore the
On Fri, Aug 22, 2014 at 11:22 AM, Sage Weil sw...@redhat.com wrote:
On Fri, 22 Aug 2014, Gregory Farnum wrote:
On Thu, Aug 21, 2014 at 3:34 PM, Sage Weil sw...@redhat.com wrote:
Sam and Josh and I discussed the state of watch/notify a couple weeks
back. Here are our notes:
http
On Fri, Aug 22, 2014 at 2:30 PM, Sage Weil sw...@redhat.com wrote:
On Fri, 22 Aug 2014, Gregory Farnum wrote:
Whereas if the notify timeout is the same time length as a watch
timeout, we can affirmatively know on a notify reply (with or without
error return codes) that every client has either
On Fri, Aug 22, 2014 at 3:59 PM, Sage Weil sw...@redhat.com wrote:
On Fri, 22 Aug 2014, Gregory Farnum wrote:
Can you give some examples of situations in which an eventual delivery
is a useful building block for something other than I know this was
delivered? I'm having trouble coming up
On Wed, Aug 20, 2014 at 3:20 PM, Christopher R. Hertel c...@redhat.com wrote:
[At end...]
- Original Message -
From: Gregory Farnum g...@inktank.com
To: Christopher R. Hertel c...@redhat.com
Cc: ceph-devel@vger.kernel.org
Sent: Tuesday, August 19, 2014 4:57:59 PM
Subject: Re: Ceph
Yeah, this is pretty weird. Can you run df on each of your OSDs and
see if any of them individually have a strange output?
-Greg
Software Engineer #42 @ http://inktank.com | http://ceph.com
On Tue, Aug 12, 2014 at 6:52 AM, Karan Singh karan.si...@csc.fi wrote:
Forgot to mention i am observing
On Wed, Aug 6, 2014 at 9:40 AM, Matt W. Benjamin m...@cohortfs.com wrote:
Hi,
We noticed that OSDs use their client_messenger, rather than their potentially
dedicated cluster_messenger, to drive MonClient. Then we found that the
Mon's setup
of the OSDMap depends on this.
Can someone
On Sun, Aug 10, 2014 at 5:22 AM, Kevin Cox kevin...@kevincox.ca wrote:
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA256
Hello,
This summer as a GSOC project I have implemented a Wireshark dissector
for Ceph. It has been committed to Wireshark master and will be
available in the next
On Thu, Aug 21, 2014 at 4:37 PM, David Zafman david.zaf...@inktank.com wrote:
Has anyone seen anything like this from an application linked with librados
using valgrind? Or a Segmentation fault on exit from such an application?
Invalid free() / delete / delete[] / realloc()
at
On Wed, Aug 13, 2014 at 4:09 AM, Viral Mehta viral@gmail.com wrote:
Hi,
I was looking at ceph/crush/builder.c crush_make_tree_bucket() function,
I would like to understand what size argument refers to ? Generally speaking
Tree Size is Total Number of the nodes. And this would be the same
On Thu, Aug 14, 2014 at 10:10 AM, Christopher R. Hertel c...@redhat.com wrote:
Ceph Folks:
Hi. I am new to Ceph but I've been around in the Open Source world for a
while, working on a variety of different projects (mostly Windows Interop
stuff). I was asked to take a look at the
On Fri, Aug 15, 2014 at 9:44 AM, Loic Dachary l...@dachary.org wrote:
Hi,
Based on Andreas suggestions for Xorbas configuration (in a previous thread)
and the last G/H discussions (
https://wiki.ceph.com/Planning/CDS/CDS_Giant_and_Hammer_%28Jun_2014%29 ),
here is a simpler configuration
On Mon, Aug 18, 2014 at 11:30 PM, Guang Yang yguan...@outlook.com wrote:
Hi ceph-devel,
David (cc’ed) reported a bug (http://tracker.ceph.com/issues/9128) which we
came across in our test cluster during our failure testing, basically the way
to reproduce it was to leave one OSD daemon down
On Tue, Aug 19, 2014 at 3:43 PM, Sage Weil sw...@redhat.com wrote:
We have had a range of bugs come up in the past because OSDs or mons have
been running different versions of the code and have encoded different
variations of the same OSDMap epoch. When two nodes in the system
disagree about
On Tue, Aug 19, 2014 at 5:32 PM, Sage Weil sw...@redhat.com wrote:
On Tue, 19 Aug 2014, Gregory Farnum wrote:
On Tue, Aug 19, 2014 at 3:43 PM, Sage Weil sw...@redhat.com wrote:
We have had a range of bugs come up in the past because OSDs or mons have
been running different versions
On Tue, Aug 19, 2014 at 9:49 PM, Sage Weil sw...@redhat.com wrote:
Right, so let's talk about how we get into that situation:
1) Our existing OSDMap is bad.
a) We were never correct
b) ...we went bad and didn't notice?
2) The Incremental we got is bad.
a) It's not the original
On Thu, Jul 31, 2014 at 1:25 AM, Sage Weil sw...@redhat.com wrote:
After the latest set of bug fixes to the FileStore file naming code I am
newly inspired to replace it with something less complex. Right now I'm
mostly thinking about HDDs, although some of this may map well onto hybrid
These sorts of questions are good for ceph-devel@vger.kernel.org,
which I've added. :)
On Thu, Jul 31, 2014 at 12:24 PM, yuelongguang fasts...@163.com wrote:
hi,all
recently i dive into the source code, i am a little confused about them,
maybe because of many threads,wait,seq.
1. what does
On Tue, Jul 29, 2014 at 7:11 PM, Sage Weil sw...@redhat.com wrote:
We've talked a bit about moving to a ~4 month (instead of 3 month)
cadence. I'm still inclined in this direction because it means fewer
stable releases that we will be maintaining and a longer and (hopefully)
more productive
On Tue, Jul 29, 2014 at 11:46 AM, Sage Weil sw...@redhat.com wrote:
On Tue, 29 Jul 2014, Wang, Zhiqiang wrote:
This fixes a bug when the time of the OSDs and clients are not synchronized
(especially when client is ahead of OSD), and the cache tier dirty ratio
reaches the threshold, the agent
On Tue, Jul 22, 2014 at 2:48 PM, Shayan Saeed shayansaee...@gmail.com wrote:
Another question along the same lines. For erasure code, same as
replicated files, the request goes through the primary member. Isn't
it possible to send the request to any of the members and get the
file. While this
On Tue, Jul 22, 2014 at 6:19 AM, Wido den Hollander w...@42on.com wrote:
Hi,
Currently on Ubuntu with Upstart when you invoke a restart like this:
$ sudo restart ceph-osd-all
It will restart all OSDs at once, which can increase the load on the system
a quite a bit.
It's better to restart
On Sun, Jul 20, 2014 at 11:33 PM, Ma, Jianpeng jianpeng...@intel.com wrote:
Hrm, I'd really like to see the startup sequence. I see the crash occurring,
but I
don't understand how it's happening — we test this pretty extensively so
there
must be something about your testing configuration
On Mon, Jul 21, 2014 at 3:27 PM, Daniel Hofmann dan...@trvx.org wrote:
Preamble: you might want to read the decent formatted version of this
mail at:
https://gist.github.com/daniel-j-h/2daae2237bb21596c97d
Motivation
--
Recently I was wondering how Ceph (i.e. Rados in particular)
On Fri, Jul 18, 2014 at 2:12 AM, Kaifeng Yao kaif...@yahoo-inc.com wrote:
I was thinking about 'PG preferred' to allow binding a PG's placement to
arbitrary OSDs. My angle is to make the PG more evenly distributed across
OSDs, thus to potentially save ~20% cost. I am searching the 'pg
wrote:
Hi Greg,
The attachment is the log.
Thanks!
-Original Message-
From: Gregory Farnum [mailto:g...@inktank.com]
Sent: Thursday, July 17, 2014 3:41 AM
To: Ma, Jianpeng
Cc: ceph-devel@vger.kernel.org
Subject: Re: [RFC][PATCH] osd: Add local_connection to fast_dispatch
On Wed, Jul 16, 2014 at 1:50 AM, James Harper ja...@ejbdigital.com.au wrote:
Can you offer some comments on what the impact is likely to be to the data in
an affected cluster? Should all data now be treated with suspicion and
restored back to before the firefly upgrade?
I am under the
On Wed, Jul 16, 2014 at 2:22 AM, Christoph Hellwig h...@infradead.org wrote:
On Tue, Jul 15, 2014 at 04:45:59PM -0700, Sage Weil wrote:
This Firefly point release fixes an potential data corruption problem
when ceph-osd daemons run on top of XFS and service Firefly librbd
clients. A recently
FYI, this sounds like an issue the userspace client (and possibly the
actual rbd class?) had as well: it looked at the HEAD parent_overlap
field even when reading from snapshots. (I don't remember if
parent_overlap is the actual parameter name, but you get the idea.)
-Greg
Software Engineer #42 @
201 - 300 of 1146 matches
Mail list logo