Re: [ceph-users] Flood of 'failed to encode map X with expected crc' on 1800 OSD cluster after upgrade

2016-07-11 Thread Wido den Hollander
> Op 12 juli 2016 om 8:47 schreef Christian Balzer : > > > > Hello, > > On Tue, 12 Jul 2016 08:39:16 +0200 (CEST) Wido den Hollander wrote: > > > Hi, > > > > I am upgrading a 1800 OSD cluster from Hammer 0.94.5 to 0.94.7 prior to > > going to Jewel and while doing so I see the monitors bein

Re: [ceph-users] Flood of 'failed to encode map X with expected crc' on 1800 OSD cluster after upgrade

2016-07-11 Thread Christian Balzer
Hello, On Tue, 12 Jul 2016 08:39:16 +0200 (CEST) Wido den Hollander wrote: > Hi, > > I am upgrading a 1800 OSD cluster from Hammer 0.94.5 to 0.94.7 prior to going > to Jewel and while doing so I see the monitors being flooded with these > messages: > Google is your friend (and so is the NSA):

[ceph-users] Flood of 'failed to encode map X with expected crc' on 1800 OSD cluster after upgrade

2016-07-11 Thread Wido den Hollander
Hi, I am upgrading a 1800 OSD cluster from Hammer 0.94.5 to 0.94.7 prior to going to Jewel and while doing so I see the monitors being flooded with these messages: 2016-07-12 08:28:12.919748 osd.1200 [WRN] failed to encode map e130549 with expected crc 2016-07-12 08:28:12.921943 osd.1338 [WRN]

Re: [ceph-users] Advice on increasing pgs

2016-07-11 Thread Christian Balzer
Hello, On Tue, 12 Jul 2016 03:43:41 + Robin Percy wrote: > First off, thanks for the great response David. > Yes, that was a very good writeup. > If I understand correctly, you're saying there are two distinct costs to > consider: peering, and backfilling. The backfilling cost is a functio

Re: [ceph-users] New to Ceph - osd autostart problem

2016-07-11 Thread Dirk Laurenz
and this- after starting the osd manually root@cephosd01:~# df Filesystem 1K-blocksUsed Available Use% Mounted on /dev/dm-0 15616412 1583180 13216900 11% / udev 10240 0 10240 0% /dev tmpfs 496564636 45020 10% /run tmpfs

[ceph-users] Repairing a broken leveldb

2016-07-11 Thread Michael Metz-Martini | SpeedPartner GmbH
Hi, while rebalancing a drive experienced read errors so I think leveldb was corrupted. Unfortunately there's currently no second copy which is up2date so I can forget this pg. Only one pg is affected (I moved all other pg's away as they had active copies on another osd. In "daily business" this

Re: [ceph-users] ceph + vmware

2016-07-11 Thread Alex Gorbachev
Hi Oliver, On Friday, July 8, 2016, Oliver Dzombic wrote: > Hi, > > does anyone have experience how to connect vmware with ceph smart ? > > iSCSI multipath does not really worked well. > NFS could be, but i think thats just too much layers in between to have > some useable performance. > > Syste

Re: [ceph-users] Mail Test

2016-07-11 Thread Ken Peng
You are welcome. But please don't send test message to a public list. :) 2016-07-12 11:07 GMT+08:00 xiongnuwang : > I have joined。 > > ___ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >

Re: [ceph-users] ceph-fuse segfaults ( jewel 10.2.2)

2016-07-11 Thread Yan, Zheng
On Tue, Jul 12, 2016 at 1:07 AM, Gregory Farnum wrote: > Oh, is this one of your custom-built packages? Are they using > tcmalloc? That difference between VSZ and RSS looks like a glibc > malloc problem. > -Greg > ceph-fuse at http://download.ceph.com/rpm-jewel/el7/x86_64/ is not linked to libtcm

Re: [ceph-users] Advice on increasing pgs

2016-07-11 Thread Robin Percy
First off, thanks for the great response David. If I understand correctly, you're saying there are two distinct costs to consider: peering, and backfilling. The backfilling cost is a function of the amount of data in our pool, and therefore won't benefit from incremental steps. But the peering cos

[ceph-users] Mail Test

2016-07-11 Thread xiongnuwang
I have joined。___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[ceph-users] Ceph v10.2.2 compile issue

2016-07-11 Thread 徐元慧
Hi All I use the Ceph stable version v10.2.2. When I begin to compile the source code, I use the make && make install. I am sure that the command make builds successfully. But the command make install would always appear the same issue for Installing /usr/local/lib/python2.7/dist-packages/ceph_det

Re: [ceph-users] Cache Tier configuration

2016-07-11 Thread Christian Balzer
Hello, On Mon, 11 Jul 2016 16:19:58 +0200 Mateusz Skała wrote: > Hello Cephers. > > Can someone help me in my cache tier configuration? I have 4 same SSD drives > 176GB (184196208K) in SSD pool, how to determine target_max_bytes? What exact SSD models are these? What version of Ceph? > I ass

Re: [ceph-users] exclusive-lock

2016-07-11 Thread Christian Balzer
Hello, In this context my first question would also be, how does one wind up with such a lock contention in the first place? And how to safely resolve this? Both of which are not Ceph problems, but those of the client stack being used or of knowledgeable, 24/7 monitoring and management. Net-spl

Re: [ceph-users] Fwd: Ceph OSD suicide himself

2016-07-11 Thread Brad Hubbard
On Mon, Jul 11, 2016 at 04:53:36PM +0200, Lionel Bouton wrote: > Le 11/07/2016 11:56, Brad Hubbard a écrit : > > On Mon, Jul 11, 2016 at 7:18 PM, Lionel Bouton > > wrote: > >> Le 11/07/2016 04:48, 한승진 a écrit : > >>> Hi cephers. > >>> > >>> I need your help for some issues. > >>> > >>> The ceph cl

Re: [ceph-users] Slow performance into windows VM

2016-07-11 Thread Christian Balzer
Hello, scrub settings will only apply to new scrubs, not running ones, as you found out. On Mon, 11 Jul 2016 15:37:49 +0300 K K wrote: > > I have tested windows instance Crystal Disk Mark. Result is: > Again, when running a test like this, check with atop/iostat how your OSDs/HDDs are doing

[ceph-users] Object creation in librbd

2016-07-11 Thread Mansour Shafaei Moghaddam
Can anyone explain or at least refer to the lines of the codes in librd by which objects are created? I need to know the relation between objects and fio's iodepth... Thanks in advance ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ce

Re: [ceph-users] New to Ceph - osd autostart problem

2016-07-11 Thread Dirk Laurenz
root@cephosd01:~# fdisk -l /dev/sdb Disk /dev/sdb: 50 GiB, 53687091200 bytes, 104857600 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 512 bytes / 512 bytes Disklabel type: gpt Disk identifier: 87B152E0-EB5D-4EB0-8FF

Re: [ceph-users] Advice on increasing pgs

2016-07-11 Thread David Turner
When you increase your PGs you're already going to be moving around all of your data. Doing a full doubling of your PGs from 64 -> 128 -> 256 -> ... -> 2048 over and over and letting it backfill to healthy every time is a lot of extra data movement that isn't needed. I would recommend setting

Re: [ceph-users] exclusive-lock

2016-07-11 Thread Jason Dillaman
Unfortunately that is correct -- the exclusive lock automatically transitions upon request in order to handle QEMU live migration. There is some on-going work to deeply integrate locking support into QEMU which would solve this live migration case and librbd could internally disable automatic lock

[ceph-users] Advice on increasing pgs

2016-07-11 Thread Robin Percy
Hello, I'm looking for some advice on how to most safely increase the pgs in our primary ceph pool. A bit of background: We're running ceph 0.80.9 and have a cluster of 126 OSDs with only 64 pgs allocated to the pool. As a result, 2 OSDs are now 88% full, while the pool is only showing as 6% used

Re: [ceph-users] ceph + vmware

2016-07-11 Thread Jake Young
I'm using this setup with ESXi 5.1 and I get very good performance. I suspect you have other issues. Reliability is another story (see Nick's posts on tgt and HA to get an idea of the awful problems you can have), but for my test labs the risk is acceptable. One change I found helpful is to run

Re: [ceph-users] Using two roots for the same pool

2016-07-11 Thread Gregory Farnum
On Mon, Jul 11, 2016 at 11:15 AM, Bob R wrote: > George, > > Check the instructions here which should allow you to test your crush rules > without applying them to your cluster. > http://dachary.org/?p=3189 > > also, fwiw, we are not using an 'emit' after each choose (note these rules > are not im

Re: [ceph-users] Using two roots for the same pool

2016-07-11 Thread Bob R
George, Check the instructions here which should allow you to test your crush rules without applying them to your cluster. http://dachary.org/?p=3189 also, fwiw, we are not using an 'emit' after each choose (note these rules are not implementing what you're trying to)- # rules rule replicated_rul

Re: [ceph-users] ceph-fuse segfaults ( jewel 10.2.2)

2016-07-11 Thread Gregory Farnum
Oh, is this one of your custom-built packages? Are they using tcmalloc? That difference between VSZ and RSS looks like a glibc malloc problem. -Greg On Mon, Jul 11, 2016 at 12:04 AM, Goncalo Borges wrote: > Hi John... > > Thank you for replying. > > Here is the result of the tests you asked but I

Re: [ceph-users] Using two roots for the same pool

2016-07-11 Thread Gregory Farnum
I'm not looking at the docs, but I think you need an "emit" statement after every choose. -Greg On Monday, July 11, 2016, George Shuklin wrote: > Hello. > > I want to try CRUSH rule with following idea: > take one OSD from root with SSD drives (and use it as primary). > take two OSD from root wi

Re: [ceph-users] ceph + vmware

2016-07-11 Thread Oliver Dzombic
Hi Mike, i was trying: https://ceph.com/dev-notes/adding-support-for-rbd-to-stgt/ ONE target, from different OSD servers directly, to multiple vmware esxi servers. A config looked like: #cat iqn.ceph-cluster_netzlaboranten-storage.conf driver iscsi bs-type rbd backing-store rbd/vmware-storag

Re: [ceph-users] Ceph OSD stuck in booting state

2016-07-11 Thread George Shuklin
Mount OSD data to temp. directory (e.g. /mnt) and check where journal pointing (ls -la /mnt/journal). It can lead to a different location from your "--osd-journal=" instruction to --mkjournal. On 07/11/2016 05:46 PM, William Josefsson wrote: Hi All, Initially, I used /dev/disk/by-partuuid/xxx

[ceph-users] Using two roots for the same pool

2016-07-11 Thread George Shuklin
Hello. I want to try CRUSH rule with following idea: take one OSD from root with SSD drives (and use it as primary). take two OSD from root with HDD drives. I've created this rule: rule rule_mix { ruleset 2 type replicated min_size 2 max_size 10 step take

Re: [ceph-users] (no subject)

2016-07-11 Thread Gaurav Goyal
Thanks! I need to create a VM having qcow2 image file as 6.7 GB but raw image as 600GB which is too big. Is there a way that i need not to convert qcow2 file to raw and it works well with rbd? Regards Gaurav Goyal On Mon, Jul 11, 2016 at 11:46 AM, Kees Meijs wrote: > Glad to hear it works now

Re: [ceph-users] New to Ceph - osd autostart problem

2016-07-11 Thread George Shuklin
Check out partition type for data partition for ceph. fdisk -l /dev/sdc On 07/11/2016 04:03 PM, Dirk Laurenz wrote: hmm, helps partially ... running /usr/sbin/ceph-disk trigger /dev/sdc1 or sdb1 works and brings osd up.. systemctl enable does not help Am 11.07.2016 um 14:49 schrieb G

Re: [ceph-users] (no subject)

2016-07-11 Thread Kees Meijs
Glad to hear it works now! Good luck with your setup. Regards, Kees On 11-07-16 17:29, Gaurav Goyal wrote: > Hello it worked for me after removing the following parameter from > /etc/nova/nova.conf file ___ ceph-users mailing list ceph-users@lists.ceph

Re: [ceph-users] ceph + vmware

2016-07-11 Thread Mike Christie
On 07/08/2016 02:22 PM, Oliver Dzombic wrote: > Hi, > > does anyone have experience how to connect vmware with ceph smart ? > > iSCSI multipath does not really worked well. Are you trying to export rbd images from multiple iscsi targets at the same time or just one target? For the HA/multiple t

Re: [ceph-users] ceph-fuse segfaults ( jewel 10.2.2)

2016-07-11 Thread Patrick Donnelly
Hi Goncalo, On Fri, Jul 8, 2016 at 3:01 AM, Goncalo Borges wrote: > 5./ I have noticed that ceph-fuse (in 10.2.2) consumes about 1.5 GB of > virtual memory when there is no applications using the filesystem. > > 7152 root 20 0 1108m 12m 5496 S 0.0 0.0 0:00.04 ceph-fuse > > When I onl

Re: [ceph-users] (no subject)

2016-07-11 Thread Gaurav Goyal
Hello it worked for me after removing the following parameter from /etc/nova/nova.conf file [root@OSKVM1 ~]# cat /etc/nova/nova.conf|grep hw_disk_discard #hw_disk_discard=unmap Though as per ceph documentation, for KILO version we must set this parameter. I am using Liberty but i am not sure if

[ceph-users] Design for Ceph Storage integration with openstack

2016-07-11 Thread Gaurav Goyal
Situation is --> I have installed openstack setup (Liberty) for my lab. Dear Ceph users, I need your suggestion for my ceph design. I have Host 1 --> Controller + Compute1 Host 2 --> Compute 2 DELL SAN storage is attached to both hosts as [root@OSKVM1 ~]# iscsiadm -m node 10.35.0.3:3260,1 iq

Re: [ceph-users] Fwd: Ceph OSD suicide himself

2016-07-11 Thread Lionel Bouton
Le 11/07/2016 11:56, Brad Hubbard a écrit : > On Mon, Jul 11, 2016 at 7:18 PM, Lionel Bouton > wrote: >> Le 11/07/2016 04:48, 한승진 a écrit : >>> Hi cephers. >>> >>> I need your help for some issues. >>> >>> The ceph cluster version is Jewel(10.2.1), and the filesytem is btrfs. >>> >>> I run 1 Mon a

Re: [ceph-users] Ceph OSD stuck in booting state

2016-07-11 Thread William Josefsson
Hi All, Initially, I used /dev/disk/by-partuuid/xxx-- when make the journal for the OSD, i.e. sudo ceph-osd -i 30 --mkjournal --osd-journal=/dev/disk/by-partuuid/2fe31ba2-1ac6-4729-9fdc-63432f50357 Then, I try to use /dev/sdx5 format and it works, i.e. sudo ceph-osd -i 30 --mkjournal

[ceph-users] Cache Tier configuration

2016-07-11 Thread Mateusz Skała
Hello Cephers. Can someone help me in my cache tier configuration? I have 4 same SSD drives 176GB (184196208K) in SSD pool, how to determine target_max_bytes? I assume that should be (4 drives* 188616916992 bytes )/ 3 replica = 251489222656 bytes *85% (because of full disk warning) It will be 213

[ceph-users] Ceph OSD stuck in booting state

2016-07-11 Thread William Josefsson
Hi Everyone, I have a problem with OSD stuck in booting state. sudo ceph daemon osd.7 status { "cluster_fsid": "724e501f-f4a3-4731-a832-c73685aabd21", "osd_fsid": "058cac6e-6c66-4eeb-865b-3d22f0e91a99", "whoami": 7, "state": "booting", "oldest_map": 1255, "newest_map": 249

Re: [ceph-users] New to Ceph - osd autostart problem

2016-07-11 Thread Dirk Laurenz
hmm, helps partially ... running /usr/sbin/ceph-disk trigger /dev/sdc1 or sdb1 works and brings osd up.. systemctl enable does not help Am 11.07.2016 um 14:49 schrieb George Shuklin: Short story how OSDs are started in systemd environments: Ceph OSD parittions has specific typecode (pa

Re: [ceph-users] OSPF to the host

2016-07-11 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Daniel Gryniewicz > Sent: 11 July 2016 13:38 > To: ceph-users@lists.ceph.com > Subject: Re: [ceph-users] OSPF to the host > > On 07/11/2016 08:23 AM, Saverio Proto wrote: > >> I'm looking at

Re: [ceph-users] drop i386 support

2016-07-11 Thread George Shuklin
On 07/11/2016 09:57 AM, kefu chai wrote: Hi Cephers, I am proposing drop the support of i386. as we don't compile Ceph with any i386 gitbuilder now[1] and hence don't test the i386 builds on sepia on a regular basis. Also, based on the assumption that people don't use i386 in production, I think

Re: [ceph-users] New to Ceph - osd autostart problem

2016-07-11 Thread George Shuklin
Short story how OSDs are started in systemd environments: Ceph OSD parittions has specific typecode (partition type 4FBD7E29-9D25-41B8-AFD0-062C0CEFF05D). It handled by udev rules shipped by ceph package: /lib/udev/rules.d/95-ceph-osd.rules It set up proper owner/group for this disk ('ceph' i

Re: [ceph-users] New to Ceph - osd autostart problem

2016-07-11 Thread Dirk Laurenz
Hi, what i do to reproduce the failure: root@cephadmin:~# ceph osd tree ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY -1 0.26340 root default -2 0.08780 host cephosd01 0 0.04390 osd.0 up 1.0 1.0 1 0.04390 osd.1 up 1.0

Re: [ceph-users] OSPF to the host

2016-07-11 Thread Daniel Gryniewicz
On 07/11/2016 08:23 AM, Saverio Proto wrote: I'm looking at the Dell S-ON switches which we can get in a Cumulus version. Any pro's and con's of using Cumulus vs old school switch OS's you may have come across? Nothing to declare here. Once configured properly the hardware works as expected

Re: [ceph-users] Slow performance into windows VM

2016-07-11 Thread K K
I have tested windows instance Crystal Disk Mark. Result is: Sequential Read : 43.049 MB/s Sequential Write : 45.181 MB/s Random Read 512KB : 78.660 MB/s Random Write 512KB : 39.292 MB/s Random Read 4KB (QD=1) : 3.511 MB/s [ 857.3 IOPS] Random Write 4KB (QD=1) : 1.380 MB/s [ 337.0 IOPS] Random Re

Re: [ceph-users] New to Ceph - osd autostart problem

2016-07-11 Thread Oliver Dzombic
Hi Dirk, without any information, its impossible to tell you anything. Please provide us some detail information about what is going wrong, including error messages and so on. As an admin you should be enough familar with your system to give us more information but just "its not working". As you

[ceph-users] New to Ceph - osd autostart problem

2016-07-11 Thread Dirk Laurenz
Hello, i'm new to ceph an try to do some first steps with ceph to understand concepts. my setup is at first completly in vm i deployed (with ceph-deploy) three monitors and three osd hosts. (3+3 vms) my frist test was to find out, if everything comes back online after a system restart

Re: [ceph-users] OSPF to the host

2016-07-11 Thread Saverio Proto
> I'm looking at the Dell S-ON switches which we can get in a Cumulus > version. Any pro's and con's of using Cumulus vs old school switch OS's you > may have come across? Nothing to declare here. Once configured properly the hardware works as expected. I never used Dell, I used switches from

Re: [ceph-users] Filestore merge and split

2016-07-11 Thread Anand Bhat
Merge happens either due to movement of objects due to CRUSH recalculation (when cluster grows or shrinks due to various reasons) or deletion of objects. Split happens when portions of objects/volumes get populated that were previously sparse. Each RADOS object by default is 4MB chunk and volumes

Re: [ceph-users] CephFS and WORM

2016-07-11 Thread Xusangdi
Thank you for the confirmation, John! As we have both CIFS&NFS users, I was wishing the feature should be implemented at the CephFS layer :< Regards, ---Sandy > -Original Message- > From: John Spray [mailto:jsp...@redhat.com] > Sent: Monday, July 11, 2016 7:28 PM > To: xusangdi 11976 (RD

Re: [ceph-users] Filestore merge and split

2016-07-11 Thread Nick Fisk
I believe splitting will happen on writes, merging I think only happens on deletions. From: Paul Renner [mailto:renner...@gmail.com] Sent: 10 July 2016 19:40 To: n...@fisk.me.uk Cc: ceph-users@lists.ceph.com Subject: Re: [ceph-users] Filestore merge and split Thanks... Do you know when s

Re: [ceph-users] ceph-fuse segfaults ( jewel 10.2.2)

2016-07-11 Thread John Spray
On Mon, Jul 11, 2016 at 8:04 AM, Goncalo Borges wrote: > Hi John... > > Thank you for replying. > > Here is the result of the tests you asked but I do not see nothing abnormal. Thanks for running through that. Yes, nothing in the output struck me as unreasonable either :-/ > Actually, your sugg

Re: [ceph-users] CephFS and WORM

2016-07-11 Thread John Spray
On Mon, Jul 11, 2016 at 9:28 AM, Xusangdi wrote: > Hi Cephers, > > > > I’m planning to set up samba/nfs based on CephFS kernel mount. The > WORM(write once read many) feature is required but I’m not > > sure if CephFS officially supports it, any suggestions? Thanks in advance. There's nothing in

[ceph-users] Misdirected clients due to kernel bug?

2016-07-11 Thread Simon Engelsman
Hello everyone, Last week, while deploying new disks in our cluster, we bump into what we believe is a kernel bug. Now everything is working fine, though we wanted to share our experience and see if other people have experienced similar behaviour. Steps we followed were: 1) First we removed DNE

Re: [ceph-users] Error EPERM when running ceph tell command

2016-07-11 Thread Andrei Mikhailovsky
Hello again Any thoughts on this issue? Cheers Andrei > From: "Andrei Mikhailovsky" > To: "ceph-users" > Sent: Wednesday, 22 June, 2016 18:02:28 > Subject: [ceph-users] Error EPERM when running ceph tell command > Hi > I am trying to run an osd level benchmark but get the following erro

Re: [ceph-users] Fwd: Ceph OSD suicide himself

2016-07-11 Thread Brad Hubbard
On Mon, Jul 11, 2016 at 7:18 PM, Lionel Bouton wrote: > Le 11/07/2016 04:48, 한승진 a écrit : >> Hi cephers. >> >> I need your help for some issues. >> >> The ceph cluster version is Jewel(10.2.1), and the filesytem is btrfs. >> >> I run 1 Mon and 48 OSD in 4 Nodes(each node has 12 OSDs). >> >> I've

Re: [ceph-users] Slow performance into windows VM

2016-07-11 Thread K K
I have change those params to all OSD and restart its: osd_scrub_end_hour = 6 osd_scrub_load_threshold = 2.5 osd_scrub_sleep = 0.1 but ceph status still show deep scrub: ceph status cluster 830beb43-9898-4fa9-98c1-ee04c1cdf69c health HEALTH_OK monmap e6: 3 mons at {block01=10.30.9.21:6789/0,obj

Re: [ceph-users] Fwd: Ceph OSD suicide himself

2016-07-11 Thread Lionel Bouton
Le 11/07/2016 04:48, 한승진 a écrit : > Hi cephers. > > I need your help for some issues. > > The ceph cluster version is Jewel(10.2.1), and the filesytem is btrfs. > > I run 1 Mon and 48 OSD in 4 Nodes(each node has 12 OSDs). > > I've experienced one of OSDs was killed himself. > > Always it issued s

Re: [ceph-users] Question about how to start ceph OSDs with systemd

2016-07-11 Thread Ernst Pijper
Hi Manual, This is a well known issue. You are definitely not the first one to hit this problem. Before Jewel i (and other as well) added the line ceph-disk activate all to /etc/rc.local to get the OSD’s running at boot. In Jewel, however, this doesn’t work anymore. Now i add these line to /et

Re: [ceph-users] Slow performance into windows VM

2016-07-11 Thread K K
Additional. OSD params aplying to scrub: "osd_scrub_invalid_stats": "true", "osd_scrub_begin_hour": "0", "osd_scrub_end_hour": "24", "osd_scrub_load_threshold": "0.5", "osd_scrub_min_interval": "86400", "osd_scrub_max_interval": "604800", "osd_scrub_interval_randomize_ratio": "0.5", "osd_scrub_chu

Re: [ceph-users] Slow performance into windows VM

2016-07-11 Thread K K
>  Still don't know your Ceph version, is it the latest Jewel? 10.1.2-0ubuntu1 > Check the ML archives, I remember people having performance issues with the 4.4 kernels. Yes, I try today to find something > These are OSD parameters, you need to query an OSD daemon.  There are: ceph --admin-daemon

Re: [ceph-users] (no subject)

2016-07-11 Thread Kees Meijs
Hi, I think there's still something misconfigured: > Invalid: 400 Bad Request: Unknown scheme 'file' found in URI (HTTP 400) It seems the RBD backend is not used as expected. Have you configured both Cinder _and_ Glance to use Ceph? Regards, Kees On 08-07-16 17:33, Gaurav Goyal wrote: > > I re

[ceph-users] CephFS and WORM

2016-07-11 Thread Xusangdi
Hi Cephers, I’m planning to set up samba/nfs based on CephFS kernel mount. The WORM(write once read many) feature is required but I’m not sure if CephFS officially supports it, any suggestions? Thanks in advance. Regards, ---Sandy

Re: [ceph-users] Slow performance into windows VM

2016-07-11 Thread Christian Balzer
Hello, On Mon, 11 Jul 2016 09:54:59 +0300 K K wrote: > > > I hope the fastest of these MONs (CPU and storage) has the lowest IP > > number and thus is the leader. > no, the lowest IP has slowest CPU. But zabbix didn't show any load at all > mons. In your use case and configuration no surprise

Re: [ceph-users] ceph-fuse segfaults ( jewel 10.2.2)

2016-07-11 Thread Goncalo Borges
On 07/11/2016 05:04 PM, Goncalo Borges wrote: Hi John... Thank you for replying. Here is the result of the tests you asked but I do not see nothing abnormal. Actually, your suggestions made me see that: 1) ceph-fuse 9.2.0 is presenting the same behaviour but with less memory consumption,

Re: [ceph-users] ceph-fuse segfaults ( jewel 10.2.2)

2016-07-11 Thread Goncalo Borges
Hi John... Thank you for replying. Here is the result of the tests you asked but I do not see nothing abnormal. Actually, your suggestions made me see that: 1) ceph-fuse 9.2.0 is presenting the same behaviour but with less memory consumption, probably, less enought so that it doesn't brake c