ceph branch status

2013-07-15 Thread ceph branch robot
-- All Branches --

Dan Mick dan.m...@inktank.com
2012-12-18 12:27:36 -0800   wip-rbd-striping
2013-07-12 14:02:18 -0700   wip-wsgi

David Zafman david.zaf...@inktank.com
2013-01-28 20:26:34 -0800   wip-wireshark-zafman
2013-03-22 18:14:10 -0700   wip-snap-test-fix

Gary Lowell gary.low...@inktank.com
2013-07-08 15:45:00 -0700   last

Gary Lowell glow...@inktank.com
2013-01-28 22:49:45 -0800   wip-3930
2013-02-05 19:29:11 -0800   wip.cppchecker
2013-02-10 22:21:52 -0800   wip-3955
2013-02-26 19:28:48 -0800   wip-system-leveldb
2013-03-01 18:55:35 -0800   wip-da-spec-1
2013-03-19 11:28:15 -0700   wip-3921
2013-04-11 23:00:05 -0700   wip-init-radosgw
2013-04-17 23:30:11 -0700   wip-4725
2013-04-21 22:06:37 -0700   wip-4752
2013-04-22 14:11:37 -0700   wip-4632
2013-05-31 11:20:40 -0700   wip-doc-prereq
2013-06-06 22:31:54 -0700   wip-build-doc
2013-07-03 17:00:31 -0700   wip-5496

Greg Farnum g...@inktank.com
2013-02-13 14:46:38 -0800   wip-mds-snap-fix
2013-02-22 19:57:53 -0800   wip-4248-snapid-journaling
2013-05-01 17:06:27 -0700   wip-optracker-4354
2013-05-31 13:28:31 -0700   wip-rgw-geo-rebase-test
2013-06-26 16:28:22 -0700   wip-rgw-geo-replica-log
2013-06-27 16:45:11 -0700   wip-rgw-geo-rebase
2013-07-01 15:43:29 -0700   wip-rgw-geo-2-replicalog3

James Page james.p...@ubuntu.com
2013-02-27 22:50:38 +   wip-debhelper-8

Joao Eduardo Luis joao.l...@inktank.com
2013-04-18 00:01:24 +0100   wip-4521-tool
2013-04-22 15:14:28 +0100   wip-4748
2013-04-24 16:42:11 +0100   wip-4521
2013-04-30 18:45:22 +0100   wip-mon-compact-dbg
2013-05-21 01:46:13 +0100   wip-monstoretool-foo
2013-05-31 16:26:02 +0100   wip-mon-cache-first-last-committed
2013-05-31 21:00:28 +0100   wip-mon-trim-b
2013-06-17 16:14:29 +0100   wip-mon-sync-fixes
2013-07-08 19:51:52 +0100   wip-rest-joao

Joe Buck jbb...@gmail.com
2013-05-02 16:32:33 -0700   wip-buck-add-terasort
2013-07-01 12:33:57 -0700   wip-rgw-geo-buck

John Wilkins john.wilk...@inktank.com
2012-12-21 15:14:37 -0800   wip-mon-docs

Josh Durgin josh.dur...@inktank.com
2013-03-01 14:45:23 -0800   wip-rbd-workunit-debug
2013-04-29 14:32:00 -0700   wip-rbd-close-image

Noah Watkins noahwatk...@gmail.com
2013-01-05 11:58:38 -0800   wip-localized-read-tests
2013-04-22 15:23:09 -0700   wip-cls-lua
2013-07-14 17:04:42 -0700   wip-osx-upstream

Roald van Loon roaldvanl...@gmail.com
2012-12-24 22:26:56 +   wip-dout

Sage Weil s...@inktank.com
2012-07-14 17:40:21 -0700   wip-osd-redirect
2012-11-30 13:47:27 -0800   wip-osd-readhole
2012-12-07 14:38:46 -0800   wip-osd-alloc
2013-01-27 11:06:08 -0800   wip-argonaut-leveldb
2013-01-29 13:46:02 -0800   wip-readdir
2013-02-11 07:05:15 -0800   wip-sim-journal-clone
2013-04-18 13:51:36 -0700   argonaut
2013-05-23 19:32:56 -0700   wip-libcephfs-rebased
2013-06-02 21:21:09 -0700   wip-fuse-bobtail
2013-06-04 22:43:04 -0700   wip-osd-push
2013-06-13 10:53:07 -0700   wip-refs
2013-06-18 17:00:00 -0700   wip-mon-refs
2013-06-21 17:59:58 -0700   wip-rgw-vstart
2013-06-24 21:23:55 -0700   bobtail
2013-06-25 13:16:45 -0700   wip-5401
2013-06-28 12:54:08 -0700   wip-mds-snap
2013-06-30 20:41:55 -0700   wip-5453
2013-07-01 17:48:09 -0700   wip-5021
2013-07-03 12:58:58 -0700   paravoid-test
2013-07-06 09:21:47 -0700   wip-mds-lazyio
2013-07-06 09:22:29 -0700   wip-mds-lazyio-cuttlefish
2013-07-06 13:00:51 -0700   wip-mds-lazyio-cuttlefish-minimal
2013-07-06 16:19:21 -0700   wip-mon-scrub-cuttlefish
2013-07-09 16:51:04 -0700   wip-scrub-paxos
2013-07-09 21:05:11 -0700   wip-5518
2013-07-10 11:03:55 -0700   wip-mon-sync
2013-07-11 12:59:56 -0700   wip-ceph-disk
2013-07-12 08:50:24 -0700   wip-libcephfs
2013-07-13 13:01:23 -0700   wip-mon-fixes
2013-07-13 14:09:10 -0700   wip-5623
2013-07-13 21:11:15 -0700   cuttlefish
2013-07-14 15:54:29 -0700   wip-4565
2013-07-14 16:20:54 -0700   wip-mon-report
2013-07-14 16:37:45 -0700   wip-rgw-warnings
2013-07-14 21:18:53 -0700   wip-5626

Sam Lang sam.l...@inktank.com
2012-11-27 15:01:58 -0600   wip-mtime-incr

Samuel Just sam.j...@inktank.com
2013-06-06 11:51:04 -0700   wip_bench_num
2013-06-06 13:08:51 -0700   wip_5238_cuttlefish
2013-06-17 14:50:53 -0700   wip-log-rewrite-sam
2013-06-19 14:54:13 -0700   wip_cuttlefish_compact_on_startup
2013-06-19 19:46:06 -0700   wip_observer
2013-07-08 14:34:50 -0700   wip-small-object-recovery

ObjectContext : check that all references are gone

2013-07-15 Thread Loic Dachary
Hi Sage,

You mentionned that there is code somewhere checking the fact that all 
references to an ObjectContext are gone and that replacing manual reference 
counting with shared_ptr should preserve this ( i.e. in relation to 
https://github.com/ceph/ceph/pull/414 replace ObjectContext pointers with 
shared_ptr ). However, I've not been able to find such code. Would you be so 
kind as to show me where it is ?

Cheers

-- 
Loïc Dachary, Artisan Logiciel Libre
All that is necessary for the triumph of evil is that good people do nothing.



signature.asc
Description: OpenPGP digital signature


OSD crash upon pool creation

2013-07-15 Thread Andrey Korolyov
Hello,

Using db2bb270e93ed44f9252d65d1d4c9b36875d0ea5 I had observed some
disaster-alike behavior after ``pool create'' command - every osd
daemon in the cluster will die at least once(some will crash times in
a row after bringing back). Please take a look on the
backtraces(almost identical) below. Issue #5637 is created in the
tracker.

Thanks!

http://xdel.ru/downloads/poolcreate.txt.gz
http://xdel.ru/downloads/poolcreate2.txt.gz
--
To unsubscribe from this list: send the line unsubscribe ceph-devel in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


OSD recovery failed because of leveldb: Corruption : checksum mismatch

2013-07-15 Thread wanhai zhu
Dear  guys :

I have a ceph cluster which is used for backend storage of kvm guest,
and this cluster has four nodes, each node has three disks.  And the
ceph version is 0.61.4.

Because of electrical power down, the ceph cluster have been shutdown
innormally several days ago. When I restarted all the nodes and
started the ceph service in each node, two osd service are down and
out, and the error message shows “ File system of the disk need to be
repair”, so I execute these CLI “xfs_check and xfs_repair -L”. After
that, I can mount the disk in the specific directory and see the raw
object data in the right state, then I start the specific osd service
but the osd service are also down and out and the error log show
“leveldb: Corruption : checksum mismatch” , because this error makes
several pg “stale+active+clean” and some pgs are lost in the cluster.

The details of the error log are as follows:

2013-07-09 16:45:31.940767 7f9a5a7ee780  0 ceph version 0.61.4
(1669132fcfc27d0c0b5e5bb93ade59d147e23404), process ceph-osd, pid 4640

2013-07-09 16:45:31.986070 7f9a5a7ee780  0 filestore(/osd0) mount
FIEMAP ioctl is supported and appears to work

2013-07-09 16:45:31.986084 7f9a5a7ee780  0 filestore(/osd0) mount
FIEMAP ioctl is disabled via 'filestore fiemap' config option

2013-07-09 16:45:31.986649 7f9a5a7ee780  0 filestore(/osd0) mount did
NOT detect btrfs

2013-07-09 16:45:32.001812 7f9a5a7ee780  0 filestore(/osd0) mount
syncfs(2) syscall fully supported (by glibc and kernel)

2013-07-09 16:45:32.001895 7f9a5a7ee780  0 filestore(/osd0) mount found snaps 

2013-07-09 16:45:32.003550 7f9a5a7ee780 -1 filestore(/osd0) Error
initializing leveldb: Corruption: checksum mismatch

2013-07-09 16:45:32.003619 7f9a5a7ee780 -1 ^[[0;31m ** ERROR: error
converting store /osd0: (1) Operation not permitted^[[0m

  In these days , I have tried several ways to resolve these
problem and recovery the osd service , but all fails and I have
exclude the cause of “xfs_check and xfs_repair” which is not
responsible for this issue. So I need your help or some advice to
resolve these problem.

  At the same time , I have some question about the ceph cluster
here, maybe someone can help me or give me a detail explanation.

1)   Are there some tools or command lines to move or recovery the
pg from one osd to another osd manually?  Or are there some ways to
fix the leveldb issue ?

2)   I used the rbd service for the guest block storage and when I
use the CLI “ceph osd pg map image-name”, I can see only one pg that
the rbd block has. Does it mean rbd block are stored in only one pg?
So does it mean the maximum of rbd block size is equal to the disk
capacity?

3)   Are there any ways or best practices to prevent the ceph
service from losing pg data when two osd services are down and out
(pool size is 2)? Customize the cluster map and rule set in order to
spilt the osd service in different failing zones as swift zone
concepts, Is that a good way?

  I need all your help and any idea or suggestion are very
appreciated.  Thanks.
--
To unsubscribe from this list: send the line unsubscribe ceph-devel in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html