Re: [ceph-users] Why choose 4k as default for stripe-width

2014-07-07 Thread Loic Dachary
Hi, Another consideration probably is the size of the packet sent over the network. Samuel Just chose this default value but I never asked about the rationale. Regarding the stripe width, it actually is stored per pool although it does not look that way. If you want to configure a pool with a

[ceph-users] Why choose 4k as default for stripe-width

2014-07-07 Thread Ma, Jianpeng
Hi Loic, At present, I test different stripe-width(4k,16k,32k,256k,512k, 1024k) with k=3/m=2. I used cosbench+librados to test the performance with 10MB-write(object size). Stripe-width avg-cpu-uitl-of-osd(user+kernel) Throghput 4k 83% 90.01

Re: [ceph-users] scrub error on firefly

2014-07-07 Thread Gregory Farnum
It's not very intuitive or easy to look at right now (there are plans from the recent developer summit to improve things), but the central log should have output about exactly what objects are busted. You'll then want to compare the copies manually to determine which ones are good or bad, get the g

[ceph-users] Connection refused after configure ceph.conf but Monitor Still Using IPV6

2014-07-07 Thread Yitao Jiang
hi, cephers i am new here , and lanunched a ceph single node with mon, mds and osds, belows is ceph.conf. I want using ipv4 with mon deamon, but it still using IPV6 *ceph.conf* --- [global] auth_service_required = none filestore_xattr_use_omap

[ceph-users] scrub error on firefly

2014-07-07 Thread Randy Smith
Greetings, I upgraded to firefly last week and I suddenly received this error: health HEALTH_ERR 1 pgs inconsistent; 1 scrub errors ceph health detail shows the following: HEALTH_ERR 1 pgs inconsistent; 1 scrub errors pg 3.c6 is active+clean+inconsistent, acting [2,5] 1 scrub errors The docs s

Re: [ceph-users] inconsistent pgs

2014-07-07 Thread Gregory Farnum
On Mon, Jul 7, 2014 at 4:39 PM, James Harper wrote: >> >> You can look at which OSDs the PGs map to. If the PGs have >> insufficient replica counts they'll report as degraded in "ceph -s" or >> "ceph -w". > > I meant in a general sense. If I have a pg that I suspect might be > insufficiently redu

Re: [ceph-users] inconsistent pgs

2014-07-07 Thread James Harper
> > You can look at which OSDs the PGs map to. If the PGs have > insufficient replica counts they'll report as degraded in "ceph -s" or > "ceph -w". I meant in a general sense. If I have a pg that I suspect might be insufficiently redundant I can look that up, but I'd like to know in advance an

Re: [ceph-users] inconsistent pgs

2014-07-07 Thread Gregory Farnum
You can look at which OSDs the PGs map to. If the PGs have insufficient replica counts they'll report as degraded in "ceph -s" or "ceph -w". Software Engineer #42 @ http://inktank.com | http://ceph.com On Mon, Jul 7, 2014 at 4:30 PM, James Harper wrote: >> >> It sounds like maybe you've got a ba

Re: [ceph-users] inconsistent pgs

2014-07-07 Thread James Harper
> > It sounds like maybe you've got a bad CRUSH map if you're seeing that. > One of the things the tunables do is make the algorithm handle a > variety of maps better, but if PGs are only mapping to one OSD you > need to fix that. > How can I tell that this is definitely the case (all copies of

Re: [ceph-users] inconsistent pgs

2014-07-07 Thread Gregory Farnum
On Mon, Jul 7, 2014 at 4:21 PM, James Harper wrote: >> >> Okay. Based on your description I think the reason for the tunables >> crashes is that either the "out" OSDs, or possibly one of the >> monitors, never got restarted. You should be able to update the >> tunables now, if you want to. (Or the

Re: [ceph-users] inconsistent pgs

2014-07-07 Thread James Harper
> > Okay. Based on your description I think the reason for the tunables > crashes is that either the "out" OSDs, or possibly one of the > monitors, never got restarted. You should be able to update the > tunables now, if you want to. (Or there's also a config option that > will disable the warning

Re: [ceph-users] inconsistent pgs

2014-07-07 Thread Gregory Farnum
Okay. Based on your description I think the reason for the tunables crashes is that either the "out" OSDs, or possibly one of the monitors, never got restarted. You should be able to update the tunables now, if you want to. (Or there's also a config option that will disable the warning; check the r

Re: [ceph-users] inconsistent pgs

2014-07-07 Thread James Harper
> > What was the exact sequence of events > Exact sequence of events was: . set retiring node OSD's out . noticed that mds's were now stuck in 'rejoining' state . messed around with restarting mds's but couldn't fix . google told me that upgrading ceph resolved such a problem for them . upgraded

Re: [ceph-users] Multipart upload on ceph 0.8 doesn't work?

2014-07-07 Thread Josh Durgin
On 07/07/2014 05:41 AM, Patrycja Szabłowska wrote: OK, the mystery is solved. From https://www.mail-archive.com/ceph-users@lists.ceph.com/msg10368.html "During a multi part upload you can't upload parts smaller than 5M" I've tried to upload smaller chunks, like 10KB. I've changed chunk size to

Re: [ceph-users] radosgw-agent failed to parse

2014-07-07 Thread Josh Durgin
On 07/04/2014 08:36 AM, Peter wrote: i am having issues running radosgw-agent to sync data between two radosgw zones. As far as i can tell both zones are running correctly. My issue is when i run the radosgw-agent command: radosgw-agent -v --src-access-key --src-secret-key --dest-access-key

[ceph-users] radosgw scalability questions

2014-07-07 Thread Fred Yang
I'm setting up federated gateway following https://ceph.com/docs/master/radosgw/federated-config/, it seems one cluster can have multiple instances serving multiple zone each(be it master or slave), but it's not clear whether I can have multiple radosgw/httpd instances in the same cluster to serve

Re: [ceph-users] Combining MDS Nodes

2014-07-07 Thread Wido den Hollander
On 07/07/2014 04:09 PM, Lazuardi Nasution wrote: Any calculation of required memory of MDS nodes related to the OSD nodes MDS memory usage depends on the amount of inodes it caches. By default this is 100k, but I'm not sure how many (kilo)bytes it uses for a single inode. But be aware, Ceph

Re: [ceph-users] Error initializing cluster client: Error

2014-07-07 Thread Pavel V. Kaygorodov
07 июля 2014 г., в 22:07, Gregory Farnum написал(а): > Do you have a ceph.conf file that the "ceph" tool can access in a > known location? Try specifying it manually with the "-c ceph.conf" Genius! -c helped! I have installed all ceph components (monitors and osds) in separate docker containe

Re: [ceph-users] Temporary degradation when adding OSD's

2014-07-07 Thread Gregory Farnum
On Mon, Jul 7, 2014 at 7:03 AM, Erik Logtenberg wrote: > Hi, > > If you add an OSD to an existing cluster, ceph will move some existing > data around so the new OSD gets its respective share of usage right away. > > Now I noticed that during this moving around, ceph reports the relevant > PG's as

Re: [ceph-users] clear active+degraded pgs

2014-07-07 Thread Gregory Farnum
CRUSH is a probabilistic algorithm. By having all those non-existent OSDs in the map, you made it so that 10/12 attempts at mapping would fail and need to be retried. CRUSH handles a lot of retries, but not enough for that to work out well. -Greg Software Engineer #42 @ http://inktank.com | http://

Re: [ceph-users] emperor -> firefly : Significant increase in RAM usage

2014-07-07 Thread Gregory Farnum
We don't test explicitly for this, but I'm surprised to hear about a jump of that magnitude. Do you have any more detailed profiling? Can you generate some? (With the tcmalloc heap dumps.) -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Mon, Jul 7, 2014 at 3:03 AM, Sylvain Mu

Re: [ceph-users] inconsistent pgs

2014-07-07 Thread Gregory Farnum
What was the exact sequence of events — were you rebalancing when you did the upgrade? Did the marked out OSDs get upgraded? Did you restart all the monitors prior to changing the tunables? (Are you *sure*?) -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Sat, Jul 5, 2014 at

Re: [ceph-users] Error initializing cluster client: Error

2014-07-07 Thread Gregory Farnum
Do you have a ceph.conf file that the "ceph" tool can access in a known location? Try specifying it manually with the "-c ceph.conf" argument. You can also add "--debug-ms 1, --debug-monc 10" and see if it outputs more useful error logs. -Greg Software Engineer #42 @ http://inktank.com | http://cep

Re: [ceph-users] Release notes for firefly not very clear wrt the tunables

2014-07-07 Thread Sage Weil
On Mon, 7 Jul 2014, James Harper wrote: > > > > Hi Sage, > > > > > Thanks for pointing this out. Is this clearer? > > > > Yes. Although it would probably be useful to say that using 'ceph osd > > crush tunables bobtail' will be enough to get rid of the warning and > > will not break compatibili

Re: [ceph-users] Release notes for firefly not very clear wrt the tunables

2014-07-07 Thread James Harper
> > Hi Sage, > > > Thanks for pointing this out. Is this clearer? > > Yes. Although it would probably be useful to say that using 'ceph osd > crush tunables bobtail' will be enough to get rid of the warning and > will not break compatibility too much (3.15 isn't that common, there > is not even

Re: [ceph-users] CEPH Cache Tiering

2014-07-07 Thread Mark Nelson
On 07/07/2014 09:17 AM, Lazuardi Nasution wrote: Hi, I'm thinking of using SSD for cache on CEPH where the SSDs are on the same OSD nodes with HDDs. My options are using CEPH Cache Tiering or using another cache software like bCache or FlashCache. On the second option, the SSDs will only cache d

Re: [ceph-users] Release notes for firefly not very clear wrt the tunables

2014-07-07 Thread Sage Weil
Hi Sylvain, Thanks for pointing this out. Is this clearer? * The default CRUSH rules and layouts are now using the 'bobtail' tunables and defaults. Upgaded clusters using the old values will now present with a health WARN state. This can be disabled by adding 'mon warn on legacy crush tu

[ceph-users] CEPH Cache Tiering

2014-07-07 Thread Lazuardi Nasution
Hi, I'm thinking of using SSD for cache on CEPH where the SSDs are on the same OSD nodes with HDDs. My options are using CEPH Cache Tiering or using another cache software like bCache or FlashCache. On the second option, the SSDs will only cache data related to HDDs on the same node only. Any exp

Re: [ceph-users] Release notes for firefly not very clear wrt the tunables

2014-07-07 Thread Sylvain Munaut
Hi Sage, > Thanks for pointing this out. Is this clearer? Yes. Although it would probably be useful to say that using 'ceph osd crush tunables bobtail' will be enough to get rid of the warning and will not break compatibility too much (3.15 isn't that common, there is not even any longterm relea

Re: [ceph-users] Combining MDS Nodes

2014-07-07 Thread Lazuardi Nasution
Any calculation of required memory of MDS nodes related to the OSD nodes total capacity? Date: Sun, 06 Jul 2014 15:00:14 +0200 > From: Wido den Hollander > To: ceph-users@lists.ceph.com > Subject: Re: [ceph-users] Combining MDS Nodes > Message-ID: <53b9485e.6070...@42on.com> > Content-Type: text/

[ceph-users] Temporary degradation when adding OSD's

2014-07-07 Thread Erik Logtenberg
Hi, If you add an OSD to an existing cluster, ceph will move some existing data around so the new OSD gets its respective share of usage right away. Now I noticed that during this moving around, ceph reports the relevant PG's as degraded. I can more or less understand the logic here: if a piece o

Re: [ceph-users] Some OSD and MDS crash

2014-07-07 Thread Pierre BLONDEAU
No chance to have those logs and even less in debug mode. I do this change 3 weeks ago. I put all my log here if it's can help : https://blondeau.users.greyc.fr/cephlog/all/ I have a chance to recover my +/- 20TB of data ? Regards Le 03/07/2014 21:48, Joao Luis a écrit : Do those logs have

Re: [ceph-users] Multipart upload on ceph 0.8 doesn't work?

2014-07-07 Thread Patrycja Szabłowska
OK, the mystery is solved. >From https://www.mail-archive.com/ceph-users@lists.ceph.com/msg10368.html "During a multi part upload you can't upload parts smaller than 5M" I've tried to upload smaller chunks, like 10KB. I've changed chunk size to 5MB and it works now. It's a pity that the Ceph's d

Re: [ceph-users] Multipart upload on ceph 0.8 doesn't work?

2014-07-07 Thread Patrycja Szabłowska
I've installed Ubuntu 12.04 in order to test multiupload with the ceph's modified fastcgi (according to this: https://ceph.com/docs/master/install/install-ceph-gateway/#apache-fastcgi-w-100-continue ). The problem is still the same: I can initiate the multi part upload or upload single part, but w

[ceph-users] Release notes for firefly not very clear wrt the tunables

2014-07-07 Thread Sylvain Munaut
Hi, In the release notes for firefly, more precisely in the "upgrading from emperor" section, you can find theses two notes : * The default CRUSH rules and layouts are now using the latest and greatest tunables and defaults. Clusters using the old values will now present with a health WARN stat

Re: [ceph-users] nginx (tengine) and radosgw

2014-07-07 Thread Sylvain Munaut
Hi, > if anyone else is looking to run radosgw without having to run apache, I > would recommend you look into tengine :) Just as a side note, you can now try the civetweb backend of rgw so you don't need fastcgi at all. We started running it that way and so far, it's working pretty good. Che

Re: [ceph-users] clear active+degraded pgs

2014-07-07 Thread hua peng
Hi, I have resolved it by run these: root@ceph2:~# ceph osd crush rm osd.0 removed item id 0 name 'osd.0' from crush map root@ceph2:~# ceph osd crush rm osd.1 removed item id 1 name 'osd.1' from crush map root@ceph2:~# ceph osd crush rm osd.2 removed item id 2 name 'osd.2' from crush map root@cep

Re: [ceph-users] nginx (tengine) and radosgw

2014-07-07 Thread Sergey Motovilovets
Hi guys. Has anyone managed to get radosgw+tengine work with horizon+nginx? I'm experiencing problems with this installation, more info here: http://www.mail-archive.com/openstack@lists.openstack.org/msg07934.html 2014-07-04 15:05 GMT+03:00 Andrei Mikhailovsky : > Hi David, > > Do you mind sha

Re: [ceph-users] emperor -> firefly : Significant increase in RAM usage

2014-07-07 Thread Sylvain Munaut
Hi, > We actually saw a decrease in memory usage after upgrading to Firefly, > though we did reboot the nodes after the upgrade while we had the > maintenance window. This is with 216 OSDs total (32-40 per node): > http://i.imgur.com/BC7RuXJ.png Interesting. Is that cluster for RBD or RGW ? My

Re: [ceph-users] error mapping device in firefly

2014-07-07 Thread Xabier Elkano
El 04/07/14 17:58, Ilya Dryomov escribió: > On Fri, Jul 4, 2014 at 11:48 AM, Xabier Elkano wrote: >> Hi, >> >> I am trying to map a rbd device in Ubuntu 14.04 (kernel 3.13.0-30-generic): >> >> # rbd -p mypool create test1 --size 500 >> >> # rbd -p mypool ls >> test1 >> >> # rbd -p mypool map test

Re: [ceph-users] emperor -> firefly : Significant increase in RAM usage

2014-07-07 Thread Dane Elwell
Hi, We actually saw a decrease in memory usage after upgrading to Firefly, though we did reboot the nodes after the upgrade while we had the maintenance window. This is with 216 OSDs total (32-40 per node): http://i.imgur.com/BC7RuXJ.png (The daily spikes were caused by a rogue updatedb process r

Re: [ceph-users] emperor -> firefly : Significant increase in RAM usage

2014-07-07 Thread Wido den Hollander
On 07/07/2014 09:28 AM, Sylvain Munaut wrote: Hi, Yesterday I finally updated our cluster to emperor (lastest stable commit) and what's fairly apparent is a much higher RAM usage on the OSD: http://i.imgur.com/qw9iKSV.png Has anyone noticed the same ? I mean 25% sudden increase in the idle ra

[ceph-users] clear active+degraded pgs

2014-07-07 Thread hua peng
Hi, I have 135 pgs degraded in the system. How can I remove them? They are in test environment, all data are not important. Thanks for the kind helps. root@ceph2:~# ceph osd tree # idweight type name up/down reweight -1 0.8398 root default -2 0.8398 host ceph2 0

Re: [ceph-users] see warning logs in ceph.log

2014-07-07 Thread hua peng
You'd have to see why the other daemons are not running, try: $ ceph osd tree And start the missing OSDs. But I have ceph osd rm them. so? Thanks. ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users

Re: [ceph-users] emperor -> firefly : Significant increase in RAM usage

2014-07-07 Thread Sylvain Munaut
Hi, > Yesterday I finally updated our cluster to emperor (lastest stable > commit) and what's fairly apparent is a much higher RAM usage on the > OSD: > > http://i.imgur.com/qw9iKSV.png > > Has anyone noticed the same ? I mean 25% sudden increase in the idle > ram usage is hard to ignore ... So

Re: [ceph-users] see warning logs in ceph.log

2014-07-07 Thread Wido den Hollander
On 07/07/2014 08:53 AM, hua peng wrote: Your PGs are not active+clean, so no I/O is possible. Are you OSDs running? $ sudo ceph -s That should give you more information about what to do. Wido Thanks. this is the info output, I saw osd is running. Can you help more? thanks. You'd have