ceph branch status
-- All Branches -- Dan Mick dan.m...@inktank.com 2012-12-18 12:27:36 -0800 wip-rbd-striping 2013-07-12 14:02:18 -0700 wip-wsgi David Zafman david.zaf...@inktank.com 2013-01-28 20:26:34 -0800 wip-wireshark-zafman 2013-03-22 18:14:10 -0700 wip-snap-test-fix Gary Lowell gary.low...@inktank.com 2013-07-08 15:45:00 -0700 last Gary Lowell glow...@inktank.com 2013-01-28 22:49:45 -0800 wip-3930 2013-02-05 19:29:11 -0800 wip.cppchecker 2013-02-10 22:21:52 -0800 wip-3955 2013-02-26 19:28:48 -0800 wip-system-leveldb 2013-03-01 18:55:35 -0800 wip-da-spec-1 2013-03-19 11:28:15 -0700 wip-3921 2013-04-11 23:00:05 -0700 wip-init-radosgw 2013-04-17 23:30:11 -0700 wip-4725 2013-04-21 22:06:37 -0700 wip-4752 2013-04-22 14:11:37 -0700 wip-4632 2013-05-31 11:20:40 -0700 wip-doc-prereq 2013-06-06 22:31:54 -0700 wip-build-doc 2013-07-03 17:00:31 -0700 wip-5496 Greg Farnum g...@inktank.com 2013-02-13 14:46:38 -0800 wip-mds-snap-fix 2013-02-22 19:57:53 -0800 wip-4248-snapid-journaling 2013-05-01 17:06:27 -0700 wip-optracker-4354 2013-05-31 13:28:31 -0700 wip-rgw-geo-rebase-test 2013-06-26 16:28:22 -0700 wip-rgw-geo-replica-log 2013-06-27 16:45:11 -0700 wip-rgw-geo-rebase 2013-07-01 15:43:29 -0700 wip-rgw-geo-2-replicalog3 James Page james.p...@ubuntu.com 2013-02-27 22:50:38 + wip-debhelper-8 Joao Eduardo Luis joao.l...@inktank.com 2013-04-18 00:01:24 +0100 wip-4521-tool 2013-04-22 15:14:28 +0100 wip-4748 2013-04-24 16:42:11 +0100 wip-4521 2013-04-30 18:45:22 +0100 wip-mon-compact-dbg 2013-05-21 01:46:13 +0100 wip-monstoretool-foo 2013-05-31 16:26:02 +0100 wip-mon-cache-first-last-committed 2013-05-31 21:00:28 +0100 wip-mon-trim-b 2013-06-17 16:14:29 +0100 wip-mon-sync-fixes 2013-07-08 19:51:52 +0100 wip-rest-joao Joe Buck jbb...@gmail.com 2013-05-02 16:32:33 -0700 wip-buck-add-terasort 2013-07-01 12:33:57 -0700 wip-rgw-geo-buck John Wilkins john.wilk...@inktank.com 2012-12-21 15:14:37 -0800 wip-mon-docs Josh Durgin josh.dur...@inktank.com 2013-03-01 14:45:23 -0800 wip-rbd-workunit-debug 2013-04-29 14:32:00 -0700 wip-rbd-close-image Noah Watkins noahwatk...@gmail.com 2013-01-05 11:58:38 -0800 wip-localized-read-tests 2013-04-22 15:23:09 -0700 wip-cls-lua 2013-07-14 17:04:42 -0700 wip-osx-upstream Roald van Loon roaldvanl...@gmail.com 2012-12-24 22:26:56 + wip-dout Sage Weil s...@inktank.com 2012-07-14 17:40:21 -0700 wip-osd-redirect 2012-11-30 13:47:27 -0800 wip-osd-readhole 2012-12-07 14:38:46 -0800 wip-osd-alloc 2013-01-27 11:06:08 -0800 wip-argonaut-leveldb 2013-01-29 13:46:02 -0800 wip-readdir 2013-02-11 07:05:15 -0800 wip-sim-journal-clone 2013-04-18 13:51:36 -0700 argonaut 2013-05-23 19:32:56 -0700 wip-libcephfs-rebased 2013-06-02 21:21:09 -0700 wip-fuse-bobtail 2013-06-04 22:43:04 -0700 wip-osd-push 2013-06-13 10:53:07 -0700 wip-refs 2013-06-18 17:00:00 -0700 wip-mon-refs 2013-06-21 17:59:58 -0700 wip-rgw-vstart 2013-06-24 21:23:55 -0700 bobtail 2013-06-25 13:16:45 -0700 wip-5401 2013-06-28 12:54:08 -0700 wip-mds-snap 2013-06-30 20:41:55 -0700 wip-5453 2013-07-01 17:48:09 -0700 wip-5021 2013-07-03 12:58:58 -0700 paravoid-test 2013-07-06 09:21:47 -0700 wip-mds-lazyio 2013-07-06 09:22:29 -0700 wip-mds-lazyio-cuttlefish 2013-07-06 13:00:51 -0700 wip-mds-lazyio-cuttlefish-minimal 2013-07-06 16:19:21 -0700 wip-mon-scrub-cuttlefish 2013-07-09 16:51:04 -0700 wip-scrub-paxos 2013-07-09 21:05:11 -0700 wip-5518 2013-07-10 11:03:55 -0700 wip-mon-sync 2013-07-11 12:59:56 -0700 wip-ceph-disk 2013-07-12 08:50:24 -0700 wip-libcephfs 2013-07-13 13:01:23 -0700 wip-mon-fixes 2013-07-13 14:09:10 -0700 wip-5623 2013-07-13 21:11:15 -0700 cuttlefish 2013-07-14 15:54:29 -0700 wip-4565 2013-07-14 16:20:54 -0700 wip-mon-report 2013-07-14 16:37:45 -0700 wip-rgw-warnings 2013-07-14 21:18:53 -0700 wip-5626 Sam Lang sam.l...@inktank.com 2012-11-27 15:01:58 -0600 wip-mtime-incr Samuel Just sam.j...@inktank.com 2013-06-06 11:51:04 -0700 wip_bench_num 2013-06-06 13:08:51 -0700 wip_5238_cuttlefish 2013-06-17 14:50:53 -0700 wip-log-rewrite-sam 2013-06-19 14:54:13 -0700 wip_cuttlefish_compact_on_startup 2013-06-19 19:46:06 -0700 wip_observer 2013-07-08 14:34:50 -0700 wip-small-object-recovery
ObjectContext : check that all references are gone
Hi Sage, You mentionned that there is code somewhere checking the fact that all references to an ObjectContext are gone and that replacing manual reference counting with shared_ptr should preserve this ( i.e. in relation to https://github.com/ceph/ceph/pull/414 replace ObjectContext pointers with shared_ptr ). However, I've not been able to find such code. Would you be so kind as to show me where it is ? Cheers -- Loïc Dachary, Artisan Logiciel Libre All that is necessary for the triumph of evil is that good people do nothing. signature.asc Description: OpenPGP digital signature
OSD crash upon pool creation
Hello, Using db2bb270e93ed44f9252d65d1d4c9b36875d0ea5 I had observed some disaster-alike behavior after ``pool create'' command - every osd daemon in the cluster will die at least once(some will crash times in a row after bringing back). Please take a look on the backtraces(almost identical) below. Issue #5637 is created in the tracker. Thanks! http://xdel.ru/downloads/poolcreate.txt.gz http://xdel.ru/downloads/poolcreate2.txt.gz -- To unsubscribe from this list: send the line unsubscribe ceph-devel in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
OSD recovery failed because of leveldb: Corruption : checksum mismatch
Dear guys : I have a ceph cluster which is used for backend storage of kvm guest, and this cluster has four nodes, each node has three disks. And the ceph version is 0.61.4. Because of electrical power down, the ceph cluster have been shutdown innormally several days ago. When I restarted all the nodes and started the ceph service in each node, two osd service are down and out, and the error message shows “ File system of the disk need to be repair”, so I execute these CLI “xfs_check and xfs_repair -L”. After that, I can mount the disk in the specific directory and see the raw object data in the right state, then I start the specific osd service but the osd service are also down and out and the error log show “leveldb: Corruption : checksum mismatch” , because this error makes several pg “stale+active+clean” and some pgs are lost in the cluster. The details of the error log are as follows: 2013-07-09 16:45:31.940767 7f9a5a7ee780 0 ceph version 0.61.4 (1669132fcfc27d0c0b5e5bb93ade59d147e23404), process ceph-osd, pid 4640 2013-07-09 16:45:31.986070 7f9a5a7ee780 0 filestore(/osd0) mount FIEMAP ioctl is supported and appears to work 2013-07-09 16:45:31.986084 7f9a5a7ee780 0 filestore(/osd0) mount FIEMAP ioctl is disabled via 'filestore fiemap' config option 2013-07-09 16:45:31.986649 7f9a5a7ee780 0 filestore(/osd0) mount did NOT detect btrfs 2013-07-09 16:45:32.001812 7f9a5a7ee780 0 filestore(/osd0) mount syncfs(2) syscall fully supported (by glibc and kernel) 2013-07-09 16:45:32.001895 7f9a5a7ee780 0 filestore(/osd0) mount found snaps 2013-07-09 16:45:32.003550 7f9a5a7ee780 -1 filestore(/osd0) Error initializing leveldb: Corruption: checksum mismatch 2013-07-09 16:45:32.003619 7f9a5a7ee780 -1 ^[[0;31m ** ERROR: error converting store /osd0: (1) Operation not permitted^[[0m In these days , I have tried several ways to resolve these problem and recovery the osd service , but all fails and I have exclude the cause of “xfs_check and xfs_repair” which is not responsible for this issue. So I need your help or some advice to resolve these problem. At the same time , I have some question about the ceph cluster here, maybe someone can help me or give me a detail explanation. 1) Are there some tools or command lines to move or recovery the pg from one osd to another osd manually? Or are there some ways to fix the leveldb issue ? 2) I used the rbd service for the guest block storage and when I use the CLI “ceph osd pg map image-name”, I can see only one pg that the rbd block has. Does it mean rbd block are stored in only one pg? So does it mean the maximum of rbd block size is equal to the disk capacity? 3) Are there any ways or best practices to prevent the ceph service from losing pg data when two osd services are down and out (pool size is 2)? Customize the cluster map and rule set in order to spilt the osd service in different failing zones as swift zone concepts, Is that a good way? I need all your help and any idea or suggestion are very appreciated. Thanks. -- To unsubscribe from this list: send the line unsubscribe ceph-devel in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html