Hi,
Is this phenomenon normal?Is there any idea about this problem?
It's a known problem with tcmalloc (search on the ceph mailing).
starting osd with TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=128M environnement
variable should help.
Another way, is to compile ceph with jemalloc instead tcmalloc
Hi!
We also observe the same behavior on our test Hammer install, and I wrote about
it some time ago:
http://permalink.gmane.org/gmane.comp.file-systems.ceph.user/22609http://permalink.gmane.org/gmane.comp.file-systems.ceph.user/22609
Jan Schremes give us some suggestions in thread, but we
Hi!
6 nodes, 70 OSDs (1-2-4Tb sata drives).
Ceph used as RBD backstore for VM images (~100VMs).
Megov Igor
CIO, Yuterra
От: Ben Hines bhi...@gmail.com
Отправлено: 14 августа 2015 г. 21:01
Кому: Межов Игорь Александрович
Копия: Voloshanenko Igor;
Hi All,
Anybody can help on this issue.
Regards
Prabu
On Mon, 17 Aug 2015 12:08:28 +0530 gjprabu lt;gjpr...@zohocorp.comgt;
wrote
Hi All,
Also please find osd information.
ceph osd dump | grep 'replicated size'
pool 2 'repo' replicated size
Hi, all,
When I do performance test with rados bench, I found tcmalloc consumed a lot
of CPU:
Samples: 265K of event 'cycles', Event count (approx.): 104385445900
+ 27.58% libtcmalloc.so.4.1.0[.]
tcmalloc::CentralFreeList::FetchFromSpans()
+ 15.25% libtcmalloc.so.4.1.0[.]
Thanks for the replies guys.
The client is set to 4MB, I haven't played with the OSD side yet as I wasn't
sure if it would make much difference, but I will give it a go. If the
client is already passing a 4MB request down through to the OSD, will it be
able to readahead any further? The next 4MB
What about https://github.com/Frontier314/EnhanceIO? Last commit 2
months ago, but no external contributors :(
The nice thing about EnhanceIO is there is no need to change device
name, unlike bcache, flashcache etc.
Best regards,
Alex
On Thu, Jul 23, 2015 at 11:02 AM, Daniel Gryniewicz
Thinking this through, pretty sure you would need to take your cluster
offline to do this. I can¹t think of a scenario where you could reliably
keep quorum as you swap your monitors to use the cluster network.
On 8/10/15, 8:59 AM, Daniel Marks daniel.ma...@codecentric.de wrote:
Hi all,
we just
Hi,
Maybe this seems like a strange question but i could not find this info in the
docs , i have following question,
For the ceph cluster you need osd daemons and monitor daemons,
On a host you can run several osd daemons (best one per drive as read in the
docs) on one host
But now my
yes. The issue is resource sharing as usual: the MONs will use disk I/O,
memory and CPU. If the cluster is small (test?) then there's no problem in
using the same disks. If the cluster starts to get bigger you may want to
dedicate resources (e.g. the disk for the MONs isn't used by an OSD). If
the
After setting up radosgw federated configuration last week and
integrating with openstack keystone auth, I have a question regarding
the configuration.
In the Keystone setup instructions for Kilo, the admin token auth
method is disabled:
How big are those OPS? Are they random? How many nodes? How many SSDs/OSDs?
What are you using to make the tests? Using atop on the OSD nodes where is
your bottleneck?
On Mon, Aug 17, 2015 at 1:05 PM, Межов Игорь Александрович me...@yuterra.ru
wrote:
Hi!
We also observe the same behavior on
I don't understand your question? You created a 1G RBD/disk and it's full.
You are able to grow it though - but that's a Linux management issue, not
ceph.
As everything is thin-provisioned you can create a RBD with an arbitrary
size - I've create one with 1PB when the cluster only had 600G/Raw
1)
~# ceph pg 5.6c7 query
Error ENOENT: i don't have pgid 5.6c7
In the osd log:
2015-08-17 16:11:45.185363 7f311be40700 0 osd.19 64706 do_command r=-2
i don't have pgid 5.6c7
2015-08-17 16:11:45.185380 7f311be40700 0 log_channel(cluster) log
[INF] : i don't have pgid 5.6c7
2) I do not
On 08/17/2015 07:03 AM, Alexandre DERUMIER wrote:
Hi,
Is this phenomenon normal?Is there any idea about this problem?
It's a known problem with tcmalloc (search on the ceph mailing).
starting osd with TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=128M environnement
variable should help.
Note
Hi, Igor.
You need to repair the PG.
for i in `ceph pg dump| grep inconsistent | grep -v 'inconsistent+repair' |
awk {'print$1'}`;do ceph pg repair $i;done
С уважением, Фасихов Ирек Нургаязович
Моб.: +79229045757
2015-08-18 8:27 GMT+03:00 Voloshanenko Igor igor.voloshane...@gmail.com:
Hi
Hi Irek, Please read careful )))
You proposal was the first, i try to do... That's why i asked about
help... (
2015-08-18 8:34 GMT+03:00 Irek Fasikhov malm...@gmail.com:
Hi, Igor.
You need to repair the PG.
for i in `ceph pg dump| grep inconsistent | grep -v 'inconsistent+repair'
| awk
Hi all, at our production cluster, due high rebalancing ((( we have 2 pgs
in inconsistent state...
root@temp:~# ceph health detail | grep inc
HEALTH_ERR 2 pgs inconsistent; 18 scrub errors
pg 2.490 is active+clean+inconsistent, acting [56,15,29]
pg 2.c4 is active+clean+inconsistent, acting
I added a couple OSD's and rebalanced, as well as added a new pool (id 10).
# ceph health detail
HEALTH_WARN 1 pgs degraded; 1 pgs stuck degraded; 5 pgs stuck unclean;
1 pgs stuck undersized; 1 pgs undersized; recovery 24379/66089446
objects misplaced (0.037%)
pg 10.4f is stuck unclean since
Hi,
Docker changed the old docker-registry project to docker-distribution
and its API to v2.
It now uses librados instead of radosgw to save data.
In some ceph installations it is easier to get access to radosgw than
to the cluster, so I've made a pull request to add radosgw support, it
would be
Many operations in the OpenStack cluster are stuck because of this. For
example, a VM cannot be removed because of operations stuck on osd.19:
2015-08-17 09:34:08.116274 7fa61e57a700 0 log_channel(cluster) log [WRN] :
slow request 1920.261825 seconds old, received at 2015-08-17
09:02:07.853997:
On Thu, Aug 13, 2015 at 1:59 PM, Adir Lev ad...@mellanox.com wrote:
Hi,
I have a CEPH cluster running on 4 physical servers, the cluster is up and
healthy
So far I was unable to connect any client to the cluster using krbd or fio
rbd plugin.
My clients can see and create images in rbd
Hi,
have a ceph cluster witch tree nodes and 32 osds.
The tree nodes have 16Gb memory but only 5Gb is in use.
Nodes are Dell Poweredge R510.
my ceph.conf:
[global]
mon_initial_members = ceph01
mon_host = 10.0.0.20,10.0.0.21,10.0.0.22
auth_cluster_required = cephx
auth_service_required =
Hi all, can you please help me with unexplained situation...
All snapshot inside ceph broken...
So, as example, we have VM template, as rbd inside ceph.
We can map it and mount to check that all ok with it
root@test:~# rbd map cold-storage/0e23c701-401d-4465-b9b4-c02939d57bb5
/dev/rbd0
Hi All,
We need to test three OSD and one image with replica 2(size 1GB). While
testing data is not writing above 1GB. Is there any option to write on third
OSD.
ceph osd pool get repo pg_num
pg_num: 126
# rbd showmapped
id pool image snap device
0 rbd
Hi All,
Also please find osd information.
ceph osd dump | grep 'replicated size'
pool 2 'repo' replicated size 2 min_size 2 crush_ruleset 0 object_hash rjenkins
pg_num 126 pgp_num 126 last_change 21573 flags hashpspool stripe_width 0
Regards
Prabu
On Mon,
26 matches
Mail list logo