Re: [ceph-users] PG overdose protection causing PG unavailability

2018-02-23 Thread David Turner
There was another part to my suggestion which was to set the initial crush weight to 0 in ceph.conf. after you add all of your osds, you could download the crush map, weight the new osds to what they should be, and upload the crush map to give them all the ability to take PGs at the same time. With

Re: [ceph-users] Ceph auth caps - make it more user error proof

2018-02-23 Thread David Turner
+1 for this. I messed up a cap on a cluster I was configuring doing this same thing. Luckily it wasn't production and I could fix it quickly. On Thu, Feb 22, 2018, 8:09 PM Gregory Farnum wrote: > On Wed, Feb 21, 2018 at 10:54 AM, Enrico Kern > wrote: > > Hey all, > > > > i would suggest some ch

Re: [ceph-users] Proper procedure to replace DB/WAL SSD

2018-02-23 Thread David Turner
Caspar, it looks like your idea should work. Worst case scenario seems like the osd wouldn't start, you'd put the old SSD back in and go back to the idea to weight them to 0, backfilling, then recreate the osds. Definitely with a try in my opinion, and I'd love to hear your experience after. Nico,

Re: [ceph-users] Ceph Bluestore performance question

2018-02-23 Thread David Turner
Your 6.7GB of DB partition for each 4TB osd is on the very small side of things. It's been discussed a few times in the ML and the general use case seems to be about 10GB DB per 1TB of osd. That would be about 40GB DB partition for each of your osds. This general rule covers most things except for

Re: [ceph-users] Some monitors have still not reached quorum

2018-02-23 Thread David Turner
The only communication on the cluster network is between osds. All other traffic for clients, mons, MDS, etc is on the public network. The cluster network is what the osds use to backfill, recover, sent replica copies of your data to the secondary osds, read parts of EC objects before the primary s

[ceph-users] wrong stretch package dependencies (was Luminous v12.2.3 released)

2018-02-23 Thread kefu chai
i am starting a new thread. replied inlined. On Thu, Feb 22, 2018 at 9:24 PM, Micha Krause wrote: > Hi, > > Debian Packages for stretch have broken dependencies: > > The following packages have unmet dependencies: > ceph-common : Depends: libleveldb1 but it is not installable >De

[ceph-users] Install previous version of Ceph

2018-02-23 Thread Scottix
Hey, We had one of our monitor servers die on us and I have a replacement computer now. In between that time you have released 12.2.3 but we are still on 12.2.2. We are on Ubuntu servers I see all the binaries are in the repo but your package cache only shows 12.2.3, is there a reason for not kee

Re: [ceph-users] Balanced MDS, all as active and recomended client settings.

2018-02-23 Thread Patrick Donnelly
On Fri, Feb 23, 2018 at 12:54 AM, Daniel Carrasco wrote: > client_permissions = false Yes, this will potentially reduce checks against the MDS. > client_quota = false This option no longer exists since Luminous; quota enforcement is no longer optional. However, if you don't have any quotas t

Re: [ceph-users] erasure coding chunk distribution

2018-02-23 Thread Gregory Farnum
On Fri, Feb 23, 2018 at 5:05 AM Dennis Benndorf < dennis.bennd...@googlemail.com> wrote: > Hi, > > at the moment we use ceph with one big rbd pool and size=4 and use a > rule to ensure that 2 copies are in each of our two rooms. This works > great for VMs. But there is some big data which should b

Re: [ceph-users] SSD Bluestore Backfills Slow

2018-02-23 Thread David Turner
Here is a [1] link to a ML thread tracking some slow backfilling on bluestore. It came down to the backfill sleep setting for them. Maybe it will help. [1] https://www.mail-archive.com/ceph-users@lists.ceph.com/msg40256.html On Fri, Feb 23, 2018 at 10:46 AM Reed Dier wrote: > Probably unrelat

Re: [ceph-users] SSD Bluestore Backfills Slow

2018-02-23 Thread Reed Dier
Probably unrelated, but I do keep seeing this odd negative objects degraded message on the fs-metadata pool: > pool fs-metadata-ssd id 16 > -34/3 objects degraded (-1133.333%) > recovery io 0 B/s, 89 keys/s, 2 objects/s > client io 51289 B/s rd, 101 kB/s wr, 0 op/s rd, 0 op/s wr Don’t mean

Re: [ceph-users] SSD Bluestore Backfills Slow

2018-02-23 Thread Reed Dier
Below is ceph -s > cluster: > id: {id} > health: HEALTH_WARN > noout flag(s) set > 260610/1068004947 objects misplaced (0.024%) > Degraded data redundancy: 23157232/1068004947 objects degraded > (2.168%), 332 pgs unclean, 328 pgs degraded, 328 pgs

Re: [ceph-users] Proper procedure to replace DB/WAL SSD

2018-02-23 Thread Nico Schottelius
A very interesting question and I would add the follow up question: Is there an easy way to add an external DB/WAL devices to an existing OSD? I suspect that it might be something on the lines of: - stop osd - create a link in ...ceph/osd/ceph-XX/block.db to the target device - (maybe run some

[ceph-users] Proper procedure to replace DB/WAL SSD

2018-02-23 Thread Caspar Smit
Hi All, What would be the proper way to preventively replace a DB/WAL SSD (when it is nearing it's DWPD/TBW limit and not failed yet). It hosts DB partitions for 5 OSD's Maybe something like: 1) ceph osd reweight 0 the 5 OSD's 2) let backfilling complete 3) destroy/remove the 5 OSD's 4) replace

Re: [ceph-users] mon service failed to start

2018-02-23 Thread Behnam Loghmani
Finally, the problem is solved by changing the whole hardware of failure server except hard disks. The last test which I have done before changing server was, cross exchanging SSD disks between failure server(node A) and one of the healthy servers(node B) and recreating the cluster. In this test w

[ceph-users] erasure coding chunk distribution

2018-02-23 Thread Dennis Benndorf
Hi, at the moment we use ceph with one big rbd pool and size=4 and use a rule to ensure that 2 copies are in each of our two rooms. This works great for VMs. But there is some big data which should be stored online but a bit cheaper. We think about using cephfs for it with erasure coding and

Re: [ceph-users] Ceph scrub logs: _scan_snaps no head for $object?

2018-02-23 Thread Mehmet
Sage Wrote( Tue, 2 Jan 2018 17:57:32 + (UTC)): Hi Stefan, Mehmet, Hi Sage, Sorry for the *extremly late* response! Are these clusters that were upgraded from prior versions, or fresh luminous installs? My Cluster was initialy installed with jewel (10.2.1) have seen some minor updates

Re: [ceph-users] Some monitors have still not reached quorum

2018-02-23 Thread Kevin Olbrich
I found a fix: It is *mandatory *to set the public network to the same network the mons use. Skipping this while the mon has another network interface, saves garbage to the monmap. - Kevin 2018-02-23 11:38 GMT+01:00 Kevin Olbrich : > I always see this: > > [mon01][DEBUG ] "mons": [ > [mon01]

Re: [ceph-users] Some monitors have still not reached quorum

2018-02-23 Thread Kevin Olbrich
I always see this: [mon01][DEBUG ] "mons": [ [mon01][DEBUG ] { [mon01][DEBUG ] "addr": "[fd91:462b:4243:47e::1:1]:6789/0", [mon01][DEBUG ] "name": "mon01", [mon01][DEBUG ] "public_addr": "[fd91:462b:4243:47e::1:1]:6789/0", [mon01][DEBUG ] "rank": 0 [mon01]

Re: [ceph-users] Many concurrent drive failures - How do I activate pgs?

2018-02-23 Thread Caspar Smit
Hi all, Thanks for all your follow ups on this. The Samsung SM863a is indeed a very good alternative, thanks! We ordered both (SM863a & DC S4600) so we can compare. Intel's response (I mean the lack of it) is not very promising. Allthough we have very good experiences with Intel DC SSD's we still

Re: [ceph-users] PG mapped to OSDs on same host although 'chooseleaf type host'

2018-02-23 Thread Wido den Hollander
On 02/23/2018 12:42 AM, Mike Lovell wrote: was the pg-upmap feature used to force a pg to get mapped to a particular osd? Yes it was. This is a semi-production cluster where the balancer module has been enabled with the upmap feature. It remapped PGs it seems to OSDs on the same host. r

[ceph-users] Some monitors have still not reached quorum

2018-02-23 Thread Kevin Olbrich
Hi! On a new cluster, I get the following error. All 3x mons are connected to the same switch and ping between them works (firewalls disabled). Mon-nodes are Ubuntu 16.04 LTS on Cep Luminous. [ceph_deploy.mon][ERROR ] Some monitors have still not reached quorum: [ceph_deploy.mon][ERROR ] mon03

Re: [ceph-users] Balanced MDS, all as active and recomended client settings.

2018-02-23 Thread Daniel Carrasco
Finally, I've changed the configuration to the following: ## ### MDS ## [mds] mds_cache_memory_limit = 792723456 mds_bal_mode = 1 ## ### Client ## [client] client_cache_size = 32768 client_mount_timeout = 30 client_oc_max_obijects = 2 client_oc_size = 1048576000 client_permissio