radosgw - bucket index sharding merge back

2014-07-07 Thread Guang Yang
Hi Yehuda, I am trying to find a way to merge back the bucket index sharding effort, and with more experience working at Ceph, I realized that the original commit was too huge which introduced trouble for review. I am thinking to break it down into multiple small commits and commit back with a

radosgw - bucket index sharding merge back

2014-07-07 Thread Guang Yang
Hi Yehuda, I am trying to find a way to merge back the bucket index sharding effort, and with more experience working at Ceph, I realized that the original commit was too huge which introduced trouble for review. I am thinking to break it down into multiple small commits and commit back with a

Re: [PATCH 09/14] libceph: introduce ceph_osdc_cancel_request()

2014-07-07 Thread Alex Elder
On 06/30/2014 09:34 AM, Ilya Dryomov wrote: On Mon, Jun 30, 2014 at 5:39 PM, Alex Elder el...@ieee.org wrote: On 06/25/2014 12:16 PM, Ilya Dryomov wrote: Introduce ceph_osdc_cancel_request() intended for canceling requests from the higher layers (rbd and cephfs). Because higher layers are in

Re: Reliability model for RADOS - effects during second failures

2014-07-07 Thread Koleos Fuscus
Hi Loic, At the risk of adding confusion to the discussion, does Indeed, you are right, answering questions with new questions adds confusion ;) I will open another thread to discuss your e-mail. I am aware that it might be difficult to answer to my previous mail but I need to understand what

ceph branch status

2014-07-07 Thread ceph branch robot
-- All Branches -- Alfredo Deza alfredo.d...@inktank.com 2013-09-27 10:33:52 -0400 wip-5900 Dan Mick dan.m...@inktank.com 2013-07-16 23:00:06 -0700 wip-5634 David Zafman david.zaf...@inktank.com 2014-06-05 00:22:34 -0700 wip-8231 Greg Farnum g...@inktank.com

Markov models for Ceph

2014-07-07 Thread Koleos Fuscus
Hello Loic, You ask previously: In other words, is there a place where one could set things like disk fail % of the time and network is X Gb/s and repairing a disk failure requires disk require reading B bytes from M disks ? As far as I understand, such factors cannot be expressed with a single

Cache tier READ_FORWARD transition

2014-07-07 Thread Luis Pabon
Hi all, I am working on OSDMonitor.cc:5325 and wanted to confirm the following read_forward cache tier transition: readforward - forward || writeback || (any num_objects_dirty == 0) forward - writeback || readforward || (any num_objects_dirty == 0) writeback - readforward ||

Re: [PATCH 10/14] rbd: rbd_obj_request_wait() should cancel the request if interrupted

2014-07-07 Thread Alex Elder
On 06/25/2014 12:16 PM, Ilya Dryomov wrote: rbd_obj_request_wait() should cancel the underlying OSD request if interrupted. Otherwise libceph will hold onto it indefinitely, causing assert failures or leaking the original object request. At first I didn't understand this. Let me see if I've

Re: Markov models for Ceph

2014-07-07 Thread Loic Dachary
Hi koleosfuscus, From http://www.kaymgee.com/Kevin_Greenan/Software_files/hfrs.tar downloaded from http://www.kaymgee.com/Kevin_Greenan/Software.html In hfrs/models/weaver_8_8_3.disk.ber.model [num states] 4 0 1 a failure 1 0 b repair 1 2 c failure 2 1 d repair 2 3 e failure [assign] a=N*lam_d

Re: Cache tier READ_FORWARD transition

2014-07-07 Thread Sage Weil
On Mon, 7 Jul 2014, Luis Pabon wrote: Hi all, I am working on OSDMonitor.cc:5325 and wanted to confirm the following read_forward cache tier transition: readforward - forward || writeback || (any num_objects_dirty == 0) forward - writeback || readforward || (any

Re: Cache tier READ_FORWARD transition

2014-07-07 Thread Mark Nelson
On 07/07/2014 02:29 PM, Sage Weil wrote: On Mon, 7 Jul 2014, Luis Pabon wrote: Hi all, I am working on OSDMonitor.cc:5325 and wanted to confirm the following read_forward cache tier transition: readforward - forward || writeback || (any num_objects_dirty == 0) forward -

Re: Cache tier READ_FORWARD transition

2014-07-07 Thread Sage Weil
On Mon, 7 Jul 2014, Mark Nelson wrote: On 07/07/2014 02:29 PM, Sage Weil wrote: On Mon, 7 Jul 2014, Luis Pabon wrote: Hi all, I am working on OSDMonitor.cc:5325 and wanted to confirm the following read_forward cache tier transition: readforward - forward ||

Re: Cache tier READ_FORWARD transition

2014-07-07 Thread Sage Weil
On Mon, 7 Jul 2014, Mark Nelson wrote: On 07/07/2014 02:29 PM, Sage Weil wrote: On Mon, 7 Jul 2014, Luis Pabon wrote: Hi all, I am working on OSDMonitor.cc:5325 and wanted to confirm the following read_forward cache tier transition: readforward - forward ||

Re: A log bug in OSD.cc

2014-07-07 Thread Gregory Farnum
Thanks! I've fixed this up in master. -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Fri, Jul 4, 2014 at 11:41 PM, 李沛伦 lpl6338...@gmail.com wrote: It seems that the 2 log sentences OSD.cc:1220 dout(10) init creating/touching snapmapper object dendl; OSD.cc:1230

Re: Cache tier READ_FORWARD transition

2014-07-07 Thread Mark Nelson
On 07/07/2014 02:43 PM, Sage Weil wrote: On Mon, 7 Jul 2014, Mark Nelson wrote: On 07/07/2014 02:29 PM, Sage Weil wrote: On Mon, 7 Jul 2014, Luis Pabon wrote: Hi all, I am working on OSDMonitor.cc:5325 and wanted to confirm the following read_forward cache tier transition:

Re: Cache tier READ_FORWARD transition

2014-07-07 Thread Luis Pabón
I think so, but I am not sure what kind of workload would benefit from that tune. Do you have any in mind? Is the reason for this tuneable value to have a more storage efficient caching tier? - Luis On 07/07/2014 03:29 PM, Sage Weil wrote: On Mon, 7 Jul 2014, Luis Pabon wrote: Hi all,

Re: Cache tier READ_FORWARD transition

2014-07-07 Thread Luis Pabón
What about the following usecase (please forgive some of my ceph architecture ignorance): If it was possible to setup OSD caching tier at the host (if the host had a dedicated SSD for accelerating I/O), then caching pools could be created to cache VM rbds, since they are inherently exclusive

Re: [PATCH 13/14] libceph: nuke ceph_osdc_unregister_linger_request()

2014-07-07 Thread Alex Elder
On 06/25/2014 12:16 PM, Ilya Dryomov wrote: Remove now unused ceph_osdc_unregister_linger_request(). Signed-off-by: Ilya Dryomov ilya.dryo...@inktank.com Looks good. Reviewed-by: Alex Elder el...@linaro.org --- include/linux/ceph/osd_client.h |2 -- net/ceph/osd_client.c |

Re: [PATCH 12/14] rbd: use rbd_obj_watch_request_helper() helper

2014-07-07 Thread Alex Elder
On 06/25/2014 12:16 PM, Ilya Dryomov wrote: Switch rbd_dev_header_{un,}watch_sync() to use the new helper and fix rbd_dev_header_unwatch_sync() to destroy watch_request structures before queuing watch-remove message while at it. This mistake slipped into commit b30a01f2a307 (rbd: fix

Re: [PATCH 11/14] rbd: add rbd_obj_watch_request_helper() helper

2014-07-07 Thread Alex Elder
On 06/25/2014 12:16 PM, Ilya Dryomov wrote: In the past, rbd_dev_header_watch_sync() used to handle both watch and unwatch requests and was entangled and leaky. Commit b30a01f2a307 (rbd: fix osd_request memory leak in __rbd_dev_header_watch_sync()) split it into two separate functions. This

Re: [PATCH 14/14] libceph: drop osd ref when canceling con work

2014-07-07 Thread Alex Elder
On 06/25/2014 12:16 PM, Ilya Dryomov wrote: queue_con() bumps osd ref count. We should do the reverse when canceling con work. Kind of unrelated to the rest of the series, but it looks good. Good to have a same-level-of-abstraction function for it as well. Reviewed-by: Alex Elder

Re: [PATCH] rbd: do not leak image_id in rbd_dev_v2_parent_info()

2014-07-07 Thread Alex Elder
On 06/30/2014 04:45 AM, Ilya Dryomov wrote: image_id is leaked if the parent happens to have been recorded already. Fix it. Looks good. Reviewed-by: Alex Elder el...@linaro.org Signed-off-by: Ilya Dryomov ilya.dryo...@inktank.com --- drivers/block/rbd.c |2 ++ 1 file changed, 2