Re: [ceph-users] tcmalloc use a lot of CPU

2015-08-17 Thread Alexandre DERUMIER
Hi, Is this phenomenon normal?Is there any idea about this problem? It's a known problem with tcmalloc (search on the ceph mailing). starting osd with TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=128M environnement variable should help. Another way, is to compile ceph with jemalloc instead tcmalloc

[ceph-users] НА: tcmalloc use a lot of CPU

2015-08-17 Thread Межов Игорь Александрович
Hi! We also observe the same behavior on our test Hammer install, and I wrote about it some time ago: http://permalink.gmane.org/gmane.comp.file-systems.ceph.user/22609http://permalink.gmane.org/gmane.comp.file-systems.ceph.user/22609 Jan Schremes give us some suggestions in thread, but we

[ceph-users] НА: НА: CEPH cache layer. Very slow

2015-08-17 Thread Межов Игорь Александрович
Hi! 6 nodes, 70 OSDs (1-2-4Tb sata drives). Ceph used as RBD backstore for VM images (~100VMs). Megov Igor CIO, Yuterra От: Ben Hines bhi...@gmail.com Отправлено: 14 августа 2015 г. 21:01 Кому: Межов Игорь Александрович Копия: Voloshanenko Igor;

Re: [ceph-users] ceph distributed osd

2015-08-17 Thread gjprabu
Hi All, Anybody can help on this issue. Regards Prabu On Mon, 17 Aug 2015 12:08:28 +0530 gjprabu lt;gjpr...@zohocorp.comgt; wrote Hi All, Also please find osd information. ceph osd dump | grep 'replicated size' pool 2 'repo' replicated size

[ceph-users] tcmalloc use a lot of CPU

2015-08-17 Thread YeYin
Hi, all, When I do performance test with rados bench, I found tcmalloc consumed a lot of CPU: Samples: 265K of event 'cycles', Event count (approx.): 104385445900 + 27.58% libtcmalloc.so.4.1.0[.] tcmalloc::CentralFreeList::FetchFromSpans() + 15.25% libtcmalloc.so.4.1.0[.]

Re: [ceph-users] How to improve single thread sequential reads?

2015-08-17 Thread Nick Fisk
Thanks for the replies guys. The client is set to 4MB, I haven't played with the OSD side yet as I wasn't sure if it would make much difference, but I will give it a go. If the client is already passing a 4MB request down through to the OSD, will it be able to readahead any further? The next 4MB

Re: [ceph-users] any recommendation of using EnhanceIO?

2015-08-17 Thread Alex Gorbachev
What about https://github.com/Frontier314/EnhanceIO? Last commit 2 months ago, but no external contributors :( The nice thing about EnhanceIO is there is no need to change device name, unlike bcache, flashcache etc. Best regards, Alex On Thu, Jul 23, 2015 at 11:02 AM, Daniel Gryniewicz

Re: [ceph-users] Is there a way to configure a cluster_network for a running cluster?

2015-08-17 Thread Will . Boege
Thinking this through, pretty sure you would need to take your cluster offline to do this. I can¹t think of a scenario where you could reliably keep quorum as you swap your monitors to use the cluster network. On 8/10/15, 8:59 AM, Daniel Marks daniel.ma...@codecentric.de wrote: Hi all, we just

[ceph-users] Question

2015-08-17 Thread Kris Vaes
Hi, Maybe this seems like a strange question but i could not find this info in the docs , i have following question, For the ceph cluster you need osd daemons and monitor daemons, On a host you can run several osd daemons (best one per drive as read in the docs) on one host But now my

Re: [ceph-users] Question

2015-08-17 Thread Luis Periquito
yes. The issue is resource sharing as usual: the MONs will use disk I/O, memory and CPU. If the cluster is small (test?) then there's no problem in using the same disks. If the cluster starts to get bigger you may want to dedicate resources (e.g. the disk for the MONs isn't used by an OSD). If the

[ceph-users] radosgw keystone integration

2015-08-17 Thread Logan V.
After setting up radosgw federated configuration last week and integrating with openstack keystone auth, I have a question regarding the configuration. In the Keystone setup instructions for Kilo, the admin token auth method is disabled:

Re: [ceph-users] НА: tcmalloc use a lot of CPU

2015-08-17 Thread Luis Periquito
How big are those OPS? Are they random? How many nodes? How many SSDs/OSDs? What are you using to make the tests? Using atop on the OSD nodes where is your bottleneck? On Mon, Aug 17, 2015 at 1:05 PM, Межов Игорь Александрович me...@yuterra.ru wrote: Hi! We also observe the same behavior on

Re: [ceph-users] ceph distributed osd

2015-08-17 Thread Luis Periquito
I don't understand your question? You created a 1G RBD/disk and it's full. You are able to grow it though - but that's a Linux management issue, not ceph. As everything is thin-provisioned you can create a RBD with an arbitrary size - I've create one with 1PB when the cluster only had 600G/Raw

Re: [ceph-users] Stuck creating pg

2015-08-17 Thread Bart Vanbrabant
1) ~# ceph pg 5.6c7 query Error ENOENT: i don't have pgid 5.6c7 In the osd log: 2015-08-17 16:11:45.185363 7f311be40700 0 osd.19 64706 do_command r=-2 i don't have pgid 5.6c7 2015-08-17 16:11:45.185380 7f311be40700 0 log_channel(cluster) log [INF] : i don't have pgid 5.6c7 2) I do not

Re: [ceph-users] tcmalloc use a lot of CPU

2015-08-17 Thread Mark Nelson
On 08/17/2015 07:03 AM, Alexandre DERUMIER wrote: Hi, Is this phenomenon normal?Is there any idea about this problem? It's a known problem with tcmalloc (search on the ceph mailing). starting osd with TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=128M environnement variable should help. Note

Re: [ceph-users] Repair inconsistent pgs..

2015-08-17 Thread Irek Fasikhov
Hi, Igor. You need to repair the PG. for i in `ceph pg dump| grep inconsistent | grep -v 'inconsistent+repair' | awk {'print$1'}`;do ceph pg repair $i;done С уважением, Фасихов Ирек Нургаязович Моб.: +79229045757 2015-08-18 8:27 GMT+03:00 Voloshanenko Igor igor.voloshane...@gmail.com: Hi

Re: [ceph-users] Repair inconsistent pgs..

2015-08-17 Thread Voloshanenko Igor
Hi Irek, Please read careful ))) You proposal was the first, i try to do... That's why i asked about help... ( 2015-08-18 8:34 GMT+03:00 Irek Fasikhov malm...@gmail.com: Hi, Igor. You need to repair the PG. for i in `ceph pg dump| grep inconsistent | grep -v 'inconsistent+repair' | awk

[ceph-users] Repair inconsistent pgs..

2015-08-17 Thread Voloshanenko Igor
Hi all, at our production cluster, due high rebalancing ((( we have 2 pgs in inconsistent state... root@temp:~# ceph health detail | grep inc HEALTH_ERR 2 pgs inconsistent; 18 scrub errors pg 2.490 is active+clean+inconsistent, acting [56,15,29] pg 2.c4 is active+clean+inconsistent, acting

Re: [ceph-users] Cluster health_warn 1 active+undersized+degraded/1 active+remapped

2015-08-17 Thread Steve Dainard
I added a couple OSD's and rebalanced, as well as added a new pool (id 10). # ceph health detail HEALTH_WARN 1 pgs degraded; 1 pgs stuck degraded; 5 pgs stuck unclean; 1 pgs stuck undersized; 1 pgs undersized; recovery 24379/66089446 objects misplaced (0.037%) pg 10.4f is stuck unclean since

[ceph-users] docker distribution

2015-08-17 Thread Lorieri
Hi, Docker changed the old docker-registry project to docker-distribution and its API to v2. It now uses librados instead of radosgw to save data. In some ceph installations it is easier to get access to radosgw than to the cluster, so I've made a pull request to add radosgw support, it would be

Re: [ceph-users] Stuck creating pg

2015-08-17 Thread Bart Vanbrabant
Many operations in the OpenStack cluster are stuck because of this. For example, a VM cannot be removed because of operations stuck on osd.19: 2015-08-17 09:34:08.116274 7fa61e57a700 0 log_channel(cluster) log [WRN] : slow request 1920.261825 seconds old, received at 2015-08-17 09:02:07.853997:

Re: [ceph-users] rbd map failed

2015-08-17 Thread Ilya Dryomov
On Thu, Aug 13, 2015 at 1:59 PM, Adir Lev ad...@mellanox.com wrote: Hi, I have a CEPH cluster running on 4 physical servers, the cluster is up and healthy So far I was unable to connect any client to the cluster using krbd or fio rbd plugin. My clients can see and create images in rbd

[ceph-users] Memory-Usage

2015-08-17 Thread Patrik Plank
Hi, have a ceph cluster witch tree nodes and 32 osds. The tree nodes have 16Gb memory but only 5Gb is in use. Nodes are Dell Poweredge R510. my ceph.conf: [global] mon_initial_members = ceph01 mon_host = 10.0.0.20,10.0.0.21,10.0.0.22 auth_cluster_required = cephx auth_service_required =

[ceph-users] Broken snapshots... CEPH 0.94.2

2015-08-17 Thread Voloshanenko Igor
Hi all, can you please help me with unexplained situation... All snapshot inside ceph broken... So, as example, we have VM template, as rbd inside ceph. We can map it and mount to check that all ok with it root@test:~# rbd map cold-storage/0e23c701-401d-4465-b9b4-c02939d57bb5 /dev/rbd0

Re: [ceph-users] ceph distributed osd

2015-08-17 Thread gjprabu
Hi All, We need to test three OSD and one image with replica 2(size 1GB). While testing data is not writing above 1GB. Is there any option to write on third OSD. ceph osd pool get repo pg_num pg_num: 126 # rbd showmapped id pool image snap device 0 rbd

Re: [ceph-users] ceph distributed osd

2015-08-17 Thread gjprabu
Hi All, Also please find osd information. ceph osd dump | grep 'replicated size' pool 2 'repo' replicated size 2 min_size 2 crush_ruleset 0 object_hash rjenkins pg_num 126 pgp_num 126 last_change 21573 flags hashpspool stripe_width 0 Regards Prabu On Mon,