Re: [ceph-users] Do you ever encountered a similar deadlock cephfs stack?

2018-10-22 Thread Yan, Zheng
> On Oct 22, 2018, at 23:06, ? ? wrote: > > > Hello: > Do you ever encountered a similar deadlock cephfs stack? > > [Sat Oct 20 15:11:40 2018] INFO: task nfsd:27191 blocked for more than 120 > seconds. > [Sat Oct 20 15:11:40 2018] Tainted: G OE > 4.14.0-49

[ceph-users] [ceph-ansible]Purging cluster using ceph-ansible stable 3.1/3.2

2018-10-22 Thread Cody
Hi folks, I tried to purge a ceph cluster using infrastructure-playbooks/purge-cluster.yml from stable 3.1 and stable 3.2 branches, but kept getting the following error immediately: ERROR! no action detected in task. This often indicates a misspelled module name, or incorrect module path. The er

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread David Turner
I don't have enough disk space on the nvme. The DB would overflow before I reached 25% utilization in the cluster. The disks are 10TB spinners and would need a minimum of 100 GB of DB space minimum based on early testing. The official docs recommend 400GB size DB for this size disk. I don't have en

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread solarflow99
Why didn't you just install the DB + WAL on the NVMe? Is this "data disk" still an ssd? On Mon, Oct 22, 2018 at 3:34 PM David Turner wrote: > And by the data disk I mean that I didn't specify a location for the DB > partition. > > On Mon, Oct 22, 2018 at 4:06 PM David Turner > wrote: > >> Tr

[ceph-users] RGW stale buckets

2018-10-22 Thread Robert Stanford
Someone deleted our rgw data pool to clean up. They recreated it afterward. This is fine in one respect, we don't need the data. But listing with radosgw-admin still shows all the buckets. How can we clean things up and get rgw to understand what actually exists, and what doesn't?

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread David Turner
No, it's exactly what I told you it was. "bluestore_bdev_partition_path" is the data path. In all of my scenarios my DB and Data are on the same partition, hence mine are the same. Your DB and WAL are on a different partition from your Data... so your DB partition is different... Whatever your m

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread Robert Stanford
That's very helpful, thanks. In your first case above your bluefs_db_partition_path and bluestore_bdev_partition path are the same. Though I have a different data and db drive, mine are different. Might this explain something? My root concern is that there is more utilization on the cluster tha

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread David Turner
My DB doesn't have a specific partition anywhere, but there's still a symlink for it to the data partition. On my home cluster with all DB, WAL, and Data on the same disk without any partitions specified there is a block symlink but no block.wal symlink. For the cluster with a specific WAL partit

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread Robert Stanford
Let me add, I have no block.wal file (which the docs suggest should be there). http://docs.ceph.com/docs/master/rados/configuration/bluestore-config-ref/ On Mon, Oct 22, 2018 at 3:13 PM Robert Stanford wrote: > > We're out of sync, I think. You have your DB on your data disk so your > block.d

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread Robert Stanford
We're out of sync, I think. You have your DB on your data disk so your block.db symlink points to that disk, right? There is however no wal symlink? So how would you verify your WAL actually lived on your NVMe? On Mon, Oct 22, 2018 at 3:07 PM David Turner wrote: > And by the data disk I mean

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread David Turner
Track down where it says they point to? Does it match what you expect? It does for me. I have my DB on my data disk and my WAL on a separate NVMe. On Mon, Oct 22, 2018 at 3:21 PM Robert Stanford wrote: > > David - is it ensured that wal and db both live where the symlink > block.db points?

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread David Turner
And by the data disk I mean that I didn't specify a location for the DB partition. On Mon, Oct 22, 2018 at 4:06 PM David Turner wrote: > Track down where it says they point to? Does it match what you expect? > It does for me. I have my DB on my data disk and my WAL on a separate NVMe. > > On M

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread Robert Stanford
David - is it ensured that wal and db both live where the symlink block.db points? I assumed that was a symlink for the db, but necessarily for the wal, because it can live in a place different than the db. On Mon, Oct 22, 2018 at 2:18 PM David Turner wrote: > You can always just go to /var/li

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread David Turner
You can always just go to /var/lib/ceph/osd/ceph-{osd-num}/ and look at where the symlinks for block and block.wal point to. On Mon, Oct 22, 2018 at 12:29 PM Robert Stanford wrote: > > That's what they say, however I did exactly this and my cluster > utilization is higher than the total pool ut

Re: [ceph-users] ceph df space usage confusion - balancing needed?

2018-10-22 Thread David Turner
I haven't had crush-compat do anything helpful for balancing my clusters. upmap has been amazing and balanced my clusters far better than anything else I've ever seen. I would go so far as to say that upmap can achieve a perfect balance. It seems to evenly distribute the PGs for each pool onto al

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread Robert Stanford
That's what they say, however I did exactly this and my cluster utilization is higher than the total pool utilization by about the number of OSDs * wal size. I want to verify that the wal is on the SSDs too but I've asked here and no one seems to know a way to verify this. Do you? Thank you, R

[ceph-users] Do you ever encountered a similar deadlock cephfs stack?

2018-10-22 Thread ? ?
Hello: Do you ever encountered a similar deadlock cephfs stack? [Sat Oct 20 15:11:40 2018] INFO: task nfsd:27191 blocked for more than 120 seconds. [Sat Oct 20 15:11:40 2018] Tainted: G OE 4.14.0-49.el7.centos.x86_64 #1 [Sat Oct 20 15:11:40 2018] "echo 0 > /proc

[ceph-users] ubuntu 16.04 failed to connect to socket /com/ubuntu/upstart connection refused

2018-10-22 Thread Liu, Changcheng
Hi all, I follow below guide to deploy ceph mimic packages on ubuntu 16.04 host. http://docs.ceph.com/docs/master/start/quick-ceph-deploy/ I always hit below problem: [DEBUG ] Setting up ceph-mon (13.2.2-1xenial) ... [DEBUG ] start: Unable to connect to Upstart: Failed to connect to

Re: [ceph-users] cephfs kernel client blocks when removing large files

2018-10-22 Thread Yan, Zheng
On Mon, Oct 22, 2018 at 7:47 PM Dylan McCulloch wrote: > > > On Mon, Oct 22, 2018 at 2:37 PM Dylan McCulloch > > wrote: > > > > > > > > > > > On Mon, Oct 22, 2018 at 9:46 AM Dylan McCulloch > > > > wrote: > > > > > > > > > > On Mon, Oct 8, 2018 at 2:57 PM Dylan McCulloch > > > > unimelb.edu.a

Re: [ceph-users] cephfs kernel client blocks when removing large files

2018-10-22 Thread Dylan McCulloch
> On Mon, Oct 22, 2018 at 2:37 PM Dylan McCulloch wrote: > > > > > > > > On Mon, Oct 22, 2018 at 9:46 AM Dylan McCulloch > > > wrote: > > > > > > > > On Mon, Oct 8, 2018 at 2:57 PM Dylan McCulloch > > > > > wrote: > > > > >> > > > > >> Hi all, > > > > >> > > > > >> > > > > >> We have identified

Re: [ceph-users] safe to remove leftover bucket index objects

2018-10-22 Thread Luis Periquito
It may be related to http://tracker.ceph.com/issues/34307 - I have a cluster whose OMAP size is larger than the stored data... On Mon, Oct 22, 2018 at 11:09 AM Wido den Hollander wrote: > > > > On 8/31/18 5:31 PM, Dan van der Ster wrote: > > So it sounds like you tried what I was going to do, and

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread Maged Mokhtar
If you specify a db on ssd and data on hdd and not explicitly specify a device for wal, wal will be placed on same ssd partition with db. Placing only wal on ssd or creating separate devices for wal and db are less common setups. /Maged On 22/10/18 09:03, Fyodor Ustinov wrote: Hi! For sha

Re: [ceph-users] safe to remove leftover bucket index objects

2018-10-22 Thread Wido den Hollander
On 8/31/18 5:31 PM, Dan van der Ster wrote: > So it sounds like you tried what I was going to do, and it broke > things. Good to know... thanks. > > In our case, what triggered the extra index objects was a user running > PUT /bucketname/ around 20 million times -- this apparently recreates > t

Re: [ceph-users] cephfs kernel client blocks when removing large files

2018-10-22 Thread Yan, Zheng
On Mon, Oct 22, 2018 at 2:37 PM Dylan McCulloch wrote: > > > > > On Mon, Oct 22, 2018 at 9:46 AM Dylan McCulloch wrote: > > > > > > On Mon, Oct 8, 2018 at 2:57 PM Dylan McCulloch > > > > wrote: > > > >> > > > >> Hi all, > > > >> > > > >> > > > >> We have identified some unexpected blocking behav

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread Fyodor Ustinov
Hi! For sharing SSD between WAL and DB what should be placed on SSD? WAL or DB? - Original Message - From: "Maged Mokhtar" To: "ceph-users" Sent: Saturday, 20 October, 2018 20:05:44 Subject: Re: [ceph-users] Drive for Wal and Db On 20/10/18 18:57, Robert Stanford wrote: Our OSDs a