Re: [ceph-users] ceph log level

2019-12-31 Thread Zhenshi Zhou
t; logarrmon[debug_mds_log]="1/5" > logarrmon[debug_mds_log_expire]="1/5" > logarrmon[debug_mds_migrator]="1/5" > logarrmon[debug_memdb]="4/5" > logarrmon[debug_mgr]="1/5" > logarrmon[debug_mgrc]="1/5" > logarrmon[debug_m

[ceph-users] ceph log level

2019-12-29 Thread Zhenshi Zhou
Hi all, OSD servers generate huge number of log. I configure 'debug_osd' to 1/5 or 1/20, but it seems not working. Is there any other option which overrides this configuration? Ceph version mimic(13.2.5) Thanks ___ ceph-users mailing list ceph-users@li

Re: [ceph-users] getting pg inconsistent periodly

2019-04-24 Thread Zhenshi Zhou
Hi, I remember that there's some bug about cephfs when upgrading from 12.2.5. Is it safe to upgrade the cluster now? Thanks Janne Johansson 于2019年4月24日周三 下午4:06写道: > > > Den ons 24 apr. 2019 kl 08:46 skrev Zhenshi Zhou : > >> Hi, >> >> I'm running a

[ceph-users] getting pg inconsistent periodly

2019-04-23 Thread Zhenshi Zhou
Hi, I'm running a cluster for a period of time. I find the cluster usually run into unhealthy state recently. With 'ceph health detail', one or two pg are inconsistent. What's more, pg in wrong state each day are not placed on the same disk, so that I don't think it's a disk problem. The cluster

[ceph-users] leak memory when mount cephfs

2019-03-19 Thread Zhenshi Zhou
Hi, I mount cephfs on my client servers. Some of the servers mount without any error whereas others don't. The error: # ceph-fuse -n client.kvm -m ceph.somedomain.com:6789 /mnt/kvm -r /kvm -d 2019-03-19 17:03:29.136 7f8c80eddc80 -1 deliberately leaking some memory 2019-03-19 17:03:29.137 7f8c80ed

Re: [ceph-users] cluster is not stable

2019-03-14 Thread Zhenshi Zhou
raken,luminous,mimic,osdmap-prune] on current monmap (epoch 3) persistent: [kraken,luminous,mimic,osdmap-prune] required: [kraken,luminous,mimic,osdmap-prune] Thanks all your helps guys:) Zhenshi Zhou 于2019年3月14日周四 下午3:20写道: > Hi, > > I'll try that command soon. >

Re: [ceph-users] cluster is not stable

2019-03-14 Thread Zhenshi Zhou
e current monmap contains no features at all, do you upgrade > cluster from luminous to mimic, > or it's a new cluster installed mimic? > > > Zhenshi Zhou 于2019年3月14日周四 下午2:37写道: > > > > Hi huang, > > > > It's a pre-production environment. If everyt

Re: [ceph-users] cluster is not stable

2019-03-13 Thread Zhenshi Zhou
ture set $f --yes-i-really-mean-it > > done > > > > If it is a production environment, you should eval the risk first, and > > maybe setup a test cluster to testing first. > > > > Zhenshi Zhou 于2019年3月14日周四 下午1:56写道: > > > > > > # ceph mon

Re: [ceph-users] cluster is not stable

2019-03-13 Thread Zhenshi Zhou
gt; min_last_epoch_clean); > > 6276 std::swap(beacon->pgs, min_last_epoch_clean_pgs); > > 6277 last_sent_beacon = now; > > 6278 } > > 6279 monc->send_mon_message(beacon); > > 6280 } else { > > 6281 dout(20) << __func__ <<

Re: [ceph-users] cluster is not stable

2019-03-13 Thread Zhenshi Zhou
osd.5 17032 send_beacon not sending huang jun 于2019年3月14日周四 下午12:30写道: > osd will not send beacons to mon if its not in ACTIVE state, > so you maybe turn on one osd's debug_osd=20 to see what is going on > > Zhenshi Zhou 于2019年3月14日周四 上午11:07写道: > > > > What's mor

Re: [ceph-users] cluster is not stable

2019-03-13 Thread Zhenshi Zhou
sd send beacon to mon, you > should also turn on debug_ms=1 on leader mon, > and restart mon process, then check the mon log to make sure mon > received osd beacon; > > Zhenshi Zhou 于2019年3月13日周三 下午8:20写道: > > > > And now, new errors are cliaming.. > > > > >

Re: [ceph-users] cluster is not stable

2019-03-13 Thread Zhenshi Zhou
And now, new errors are cliaming.. [image: image.png] Zhenshi Zhou 于2019年3月13日周三 下午2:58写道: > Hi, > > I didn't set osd_beacon_report_interval as it must be the default value. > I have set osd_beacon_report_interval to 60 and debug_mon to 10. > > Attachment is the leader

Re: [ceph-users] mount cephfs on ceph servers

2019-03-13 Thread Zhenshi Zhou
> Thank you. > >> > >> > >> > >> Daniele > >> > >> > >> > >> On 06/03/19 16:40, Jake Grimmett wrote: > >> > >>> Just to add "+1" on this datapoint, based on one month usage on > Mimic > >

Re: [ceph-users] cluster is not stable

2019-03-12 Thread Zhenshi Zhou
our mons > from all your OSD servers? > > I’ve seen this where a route wasn’t working one direction, so it made OSDs > flap when it used that mon to check availability: > > On Wed, 13 Mar 2019 at 11:50 AM, Zhenshi Zhou wrote: > >> After checking the network and syslog/dmsg

Re: [ceph-users] cluster is not stable

2019-03-12 Thread Zhenshi Zhou
d one host in a test cluster. >> >> Kevin >> >> >> Am Di., 12. März 2019 um 09:31 Uhr schrieb Zhenshi Zhou < >> deader...@gmail.com>: >> >>> Hi, >>> >>> I deployed a ceph cluster with good performance. But the logs >

Re: [ceph-users] cluster is not stable

2019-03-12 Thread Zhenshi Zhou
eturn to the normal state. Thanks. Kevin Olbrich 于2019年3月12日周二 下午5:44写道: > Are you sure that firewalld is stopped and disabled? > Looks exactly like that when I missed one host in a test cluster. > > Kevin > > > Am Di., 12. März 2019 um 09:31 Uhr schrieb Zhenshi Zhou < &g

Re: [ceph-users] cluster is not stable

2019-03-12 Thread Zhenshi Zhou
> dmesg, switches etc. for hints that a network interruption may have > occured. > > Regards, > Eugen > > > Zitat von Zhenshi Zhou : > > > Hi, > > > > I deployed a ceph cluster with good performance. But the logs > > indicate that the cluster is not a

[ceph-users] cluster is not stable

2019-03-12 Thread Zhenshi Zhou
Hi, I deployed a ceph cluster with good performance. But the logs indicate that the cluster is not as stable as I think it should be. The log shows the monitors mark some osd as down periodly: [image: image.png] I didn't find any useful information in osd logs. ceph version 13.2.4 mimic (stable

[ceph-users] mount cephfs on ceph servers

2019-03-05 Thread Zhenshi Zhou
Hi, I'm gonna mount cephfs from my ceph servers for some reason, including monitors, metadata servers and osd servers. I know it's not a best practice. But what is the exact potential danger if I mount cephfs from its own server? Thanks ___ ceph-users m

Re: [ceph-users] ceph migration

2019-02-27 Thread Zhenshi Zhou
Hi, The servers have moved to the new datacenter and I got it online following the instruction. # ceph -s cluster: id: 7712ab7e-3c38-44b3-96d3-4e1de9da0ff6 health: HEALTH_OK services: mon: 3 daemons, quorum ceph-mon1,ceph-mon2,ceph-mon3 mgr: ceph-mon3(active), standbys: c

Re: [ceph-users] ceph migration

2019-02-25 Thread Zhenshi Zhou
isible in the databases, and those will not > change. > > I'm not sure where old IPs will be still visible, could you clarify > that, please? > > Regards, > Eugen > > > [1] http://docs.ceph.com/docs/luminous/rados/operations/add-or-rm-mons/ > > > Zitat von

[ceph-users] ceph migration

2019-02-25 Thread Zhenshi Zhou
Hi, I deployed a new cluster(mimic). Now I have to move all servers in this cluster to another place, with new IP. I'm not sure if the cluster will run well or not after I modify config files, include /etc/hosts and /etc/ceph/ceph.conf. Fortunately, the cluster has no data at present. I never en

Re: [ceph-users] create osd failed due to cephx authentication

2019-01-24 Thread Zhenshi Zhou
27;allow profile osd' > mgr 'allow profile osd' -i /var/lib/ceph/osd/ceph-15/keyring > > ceph osd create > > sudo -u ceph ceph osd crush add osd.15 0.4 host=c04 > > systemctl start ceph-osd@15 > systemctl enable ceph-osd@15 > > > > > > -Ori

[ceph-users] create osd failed due to cephx authentication

2019-01-24 Thread Zhenshi Zhou
Hi, I'm installing a new ceph cluster manually. I get errors when I create osd: # ceph-osd -i 0 --mkfs --mkkey 2019-01-24 17:07:44.045 7f45f497b1c0 -1 auth: unable to find a keyring on /var/lib/ceph/osd/ceph-0/keyring: (2) No such file or directory 2019-01-24 17:07:44.045 7f45f497b1c0 -1 monclien

Re: [ceph-users] cephfs client operation record

2019-01-01 Thread Zhenshi Zhou
Thanks for the reply Zheng. Yan, Zheng 于2019年1月2日周三 下午3:08写道: > On Wed, Jan 2, 2019 at 11:12 AM Zhenshi Zhou wrote: > > > > Hi all, > > > > I have a cluster on Luminous(12.2.8). > > Is there a way I can check clients' operation record

[ceph-users] cephfs client operation record

2019-01-01 Thread Zhenshi Zhou
Hi all, I have a cluster on Luminous(12.2.8). Is there a way I can check clients' operation records? Thanks ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Re: [ceph-users] ceph remote disaster recovery plan

2018-12-17 Thread Zhenshi Zhou
月18日周二 上午8:45写道: > On Thu, Dec 13, 2018 at 5:01 AM Zhenshi Zhou wrote: > > > > Hi all, > > > > I'm running a luminous cluster with tens of OSDs and > > the cluster runs well. As the data grows, ceph becomes > > more and more important. > > > >

Re: [ceph-users] ceph remote disaster recovery plan

2018-12-17 Thread Zhenshi Zhou
Any idea would be appreciated:) Zhenshi Zhou 于2018年12月13日周四 下午6:01写道: > Hi all, > > I'm running a luminous cluster with tens of OSDs and > the cluster runs well. As the data grows, ceph becomes > more and more important. > > What worries me is that many services wil

[ceph-users] ceph remote disaster recovery plan

2018-12-13 Thread Zhenshi Zhou
Hi all, I'm running a luminous cluster with tens of OSDs and the cluster runs well. As the data grows, ceph becomes more and more important. What worries me is that many services will down if the cluster is out, for instance, the engine room is out of electric or all ceph node are down at the sam

Re: [ceph-users] move directories in cephfs

2018-12-12 Thread Zhenshi Zhou
Hi Thanks for the explanation. I did a test few moments ago. Everything goes just like what I expect. Thanks for your helps :) Konstantin Shalygin 于2018年12月12日周三 下午4:57写道: > Hi > > Than means, the 'mv' operation should be done if src and dst > are in the same pool, and the client should have

Re: [ceph-users] move directories in cephfs

2018-12-11 Thread Zhenshi Zhou
symlink to the old pool. > > > > > > > > > > -Original Message- > > From: Jack [mailto:c...@jack.fr.eu.org] > > Sent: 10 December 2018 15:14 > > To: ceph-users@lists.ceph.com > > Subject: Re: [ceph-users] move directories in cephfs > > &g

Re: [ceph-users] move directories in cephfs

2018-12-10 Thread Zhenshi Zhou
move directories in cephfs > > Having the / mounted somewhere, you can simply "mv" directories around > > On 12/10/2018 02:59 PM, Zhenshi Zhou wrote: > > Hi, > > > > Is there a way I can move sub-directories outside the directory. > > For instance, a directory

Re: [ceph-users] move directories in cephfs

2018-12-10 Thread Zhenshi Zhou
the / mounted somewhere, you can simply "mv" directories around > > On 12/10/2018 02:59 PM, Zhenshi Zhou wrote: > > Hi, > > > > Is there a way I can move sub-directories outside the directory. > > For instance, a directory /parent contains 3 sub-directories

[ceph-users] move directories in cephfs

2018-12-10 Thread Zhenshi Zhou
Hi, Is there a way I can move sub-directories outside the directory. For instance, a directory /parent contains 3 sub-directories /parent/a, /parent/b, /parent/c. All these directories have huge data in it. I'm gonna move /parent/b to /b. I don't want to copy the whole directory outside cause it w

Re: [ceph-users] 【cephfs】cephfs hung when scp/rsync large files

2018-12-05 Thread Zhenshi Zhou
Hi I have the same issue a few months ago. One of my client hung on waiting for a file writing. And other clients seems not being effected by it. However, if other clients access to the same hanged direcotry, it would hang there as well. My cluster is 12.2.8 and I use kernel client on other serve

Re: [ceph-users] how to mount one of the cephfs namespace using ceph-fuse?

2018-12-04 Thread Zhenshi Zhou
Hi I can use this mount cephfs manually. But how to edit fstab so that the system will auto-mount cephfs by ceph-fuse? Thanks Yan, Zheng 于2018年11月20日周二 下午8:08写道: > ceph-fuse --client_mds_namespace=xxx > On Tue, Nov 20, 2018 at 7:33 PM ST Wong (ITSC) wrote: > > > > Hi all, > > > > > > > > We’r

Re: [ceph-users] install ceph-fuse on centos5

2018-11-30 Thread Zhenshi Zhou
Does anybody know how to install ceph-fuse on Centos5? Zhenshi Zhou 于2018年11月29日周四 下午4:50写道: > Hi, > > I have a Centos5 server with kernel version 2.6.18. > Does it support to mount cephfs with ceph-fuse? > > Thanks > ___ ceph-us

Re: [ceph-users] client failing to respond to cache pressure

2018-11-29 Thread Zhenshi Zhou
Hi The cluster recovers healthy after I set "mds_cache_memory_limit" from 4G to 8G. Zhenshi Zhou 于2018年11月30日周五 上午11:04写道: > Hi > > I used to get warning message claims client faling to respond > to cache pressure. After I switch rockdb and wal data to ssd, the >

[ceph-users] client failing to respond to cache pressure

2018-11-29 Thread Zhenshi Zhou
Hi I used to get warning message claims client faling to respond to cache pressure. After I switch rockdb and wal data to ssd, the message seems disappeared. However it shows again yesterday and the message looks a little different: *MDS_CLIENT_RECALL_MANY: 1 MDSs have many * *clients failing to

[ceph-users] install ceph-fuse on centos5

2018-11-29 Thread Zhenshi Zhou
Hi, I have a Centos5 server with kernel version 2.6.18. Does it support to mount cephfs with ceph-fuse? Thanks ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Re: [ceph-users] get cephfs mounting clients' infomation

2018-11-19 Thread Zhenshi Zhou
Hi Yan, I can get the usage of sub dirctory on the client side. Is there a way I can get it from the server? Thanks. Yan, Zheng 于2018年11月19日周一 下午3:08写道: > On Mon, Nov 19, 2018 at 3:06 PM Zhenshi Zhou wrote: > > > > Many thanks Yan! > > > > This command can get I

Re: [ceph-users] get cephfs mounting clients' infomation

2018-11-18 Thread Zhenshi Zhou
tc. I wanna know how much space is taken up by each of them. Thanks. Yan, Zheng 于2018年11月19日周一 下午2:50写道: > 'ceph daemon mds.xx session ls' > On Mon, Nov 19, 2018 at 2:40 PM Zhenshi Zhou wrote: > > > > Hi, > > > > I have a cluster providing cephfs and it looks

[ceph-users] get cephfs mounting clients' infomation

2018-11-18 Thread Zhenshi Zhou
Hi, I have a cluster providing cephfs and it looks well. But as times goes by, more and more clients use it. I wanna write a script for getting the clients' informations so that I can keep everything in good order. I google a lot but dont find any solution which I can get clients information. Is

Re: [ceph-users] Migration osds to Bluestore on Ubuntu 14.04 Trusty

2018-11-16 Thread Zhenshi Zhou
Hi Klimenko, I did a migration from filestore to bluestroe on centos7 with ceph version 12.2.5. As it's the pro environment, I removed and recreated OSDs on each server at a time, online. Although I migreated on centos, I create osd manually so that you can have a try. Except one raid1 disk for s

Re: [ceph-users] upgrade ceph from L to M

2018-11-13 Thread Zhenshi Zhou
Thanks for the reply Wido. I will delay the upgrading plan. Wido den Hollander 于2018年11月13日周二 下午10:35写道: > > > On 11/13/18 12:49 PM, Zhenshi Zhou wrote: > > Hi > > > > I remember that there was a bug when using cephfs after > > upgrading ceph from L to M.

[ceph-users] upgrade ceph from L to M

2018-11-13 Thread Zhenshi Zhou
Hi I remember that there was a bug when using cephfs after upgrading ceph from L to M. Is that bug fixed now? ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Re: [ceph-users] cephfs quota limit

2018-11-07 Thread Zhenshi Zhou
:57:48PM +0800, Zhenshi Zhou wrote: > > Hi, > > I'm wondering whether cephfs have quota limit options. > > I use kernel client and ceph version is 12.2.8. > > Thanks > CephFS has quota support, see > http://docs.ceph.com/docs/luminous/cephfs/quota/. > Th

[ceph-users] cephfs quota limit

2018-11-06 Thread Zhenshi Zhou
Hi, I'm wondering whether cephfs have quota limit options. I use kernel client and ceph version is 12.2.8. Thanks ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Re: [ceph-users] Ceph osd logs

2018-10-18 Thread Zhenshi Zhou
> > On Mon, Oct 15, 2018 at 6:00 AM Zhenshi Zhou wrote: > >> Hi, >> >> I added some OSDs into cluster(luminous) lately. The osds use >> bluestoreand everything goes fine. But there is no osd log in the >> log file. The log directory has only empty fi

Re: [ceph-users] ceph pg/pgp number calculation

2018-10-18 Thread Zhenshi Zhou
> > On Tue, Oct 16, 2018, 3:35 AM Zhenshi Zhou wrote: > >> Hi, >> >> I have a cluster serving rbd and cephfs storage for a period of >> time. I added rgw in the cluster yesterday and wanted it to server >> object storage. Everything seems good. >> >

[ceph-users] ceph pg/pgp number calculation

2018-10-16 Thread Zhenshi Zhou
Hi, I have a cluster serving rbd and cephfs storage for a period of time. I added rgw in the cluster yesterday and wanted it to server object storage. Everything seems good. What I'm confused is how to calculate the pg/pgp number. As we all know, the formula of calculating pgs is: Total PGs = ((

[ceph-users] Ceph osd logs

2018-10-14 Thread Zhenshi Zhou
Hi, I added some OSDs into cluster(luminous) lately. The osds use bluestoreand everything goes fine. But there is no osd log in the log file. The log directory has only empty files. I check my settings, "ceph daemon osd.x config show", and I get "debug_osd": "1/5". How can I get the new osds' lo

Re: [ceph-users] No space left on device

2018-09-26 Thread Zhenshi Zhou
Answer to myself, What I get wrong is that the file number is much more than 2. My db shows the directory has 52 files so that it alarmed "no space left". I solve this by increasing "mds_bal_fragment_size_max" to 100. Thanks Zhenshi Zhou 于2018年9月26日周三 下午

[ceph-users] No space left on device

2018-09-26 Thread Zhenshi Zhou
Hi, I encountered an issue similar to bug 19438 . I have an attachment directory in cephfs and one of the sub directory includes about 20-30 thousand small files from 100k-1M each. When I created file in this sub directory it alarms: "cannot touch ‘attachmen

Re: [ceph-users] cephfs is growing up rapidly

2018-09-14 Thread Zhenshi Zhou
"recovery_completed": 7, "ireq_enqueue_scrub": 0, "ireq_exportdir": 0, "ireq_flush": 0, "ireq_fragmentdir": 0, "ireq_fragstats": 0, "ireq_inodestats": 0 } "purge_queue

[ceph-users] cephfs is growing up rapidly

2018-09-13 Thread Zhenshi Zhou
Hi, I have a ceph cluster of version 12.2.5 on centos7. I created 3 pools, 'rbd' for rbd storage, as well as 'cephfs_data' and 'cephfs_meta' for cephfs. Cephfs is used for backing up by rsync and volumes mounting by docker. The size of backup files is 3.5T. Besides, docker use less than 60G spac

Re: [ceph-users] failing to respond to cache pressure

2018-09-07 Thread Zhenshi Zhou
gt; > ceph@host1:~ $ ceph daemon mds.host1 config set client_oc_size 104857600 > > > > Although I still encountered one pressure message recently the total > > amount of these messages has decreased significantly. > > > > Regards, > > Eugen > > > > > &

Re: [ceph-users] cephfs kernel client hangs

2018-08-27 Thread Zhenshi Zhou
28 use. > On Sat, Aug 25, 2018 at 6:12 AM Zhenshi Zhou wrote: > > > > Hi, > > This time, osdc: > > > > REQUESTS 0 homeless 0 > > LINGER REQUESTS > > > > monc: > > > > have monmap 2 want 3+ > > have osdmap 4545 want 45

[ceph-users] Can I deploy wal and db of more than one osd in one partition

2018-08-26 Thread Zhenshi Zhou
Hi, I have 4 osd nodes with 4 hdd and 1 ssd on each. I'm gonna add these osds in an existing cluster. What I'm confused is that how to deal with the ssd. Can I deploy 4 osd with wal and db in one ssd partition such as: # ceph-disk prepare --bluestore --block.db /dev/sdc --block.wal /dev/sdc /dev/sd

Re: [ceph-users] cephfs kernel client hangs

2018-08-25 Thread Zhenshi Zhou
ere hang request in /sys/kernel/debug/ceph//osdc > > On Fri, Aug 24, 2018 at 9:32 PM Zhenshi Zhou wrote: > > > > I'm afaid that the client hangs again...the log shows: > > > > 2018-08-24 21:27:54.714334 [WRN] slow request 62.607608 seconds old, > received at 2

Re: [ceph-users] cephfs kernel client hangs

2018-08-24 Thread Zhenshi Zhou
s report slow requests (MDS_SLOW_REQUEST) 2018-08-24 21:14:28.167096 [WRN] MDS health message (mds.0): 1 slow requests are blocked > 30 sec Yan, Zheng 于2018年8月14日周二 下午3:13写道: > On Mon, Aug 13, 2018 at 9:55 PM Zhenshi Zhou wrote: > > > > Hi Burkhard, > > I'm s

Re: [ceph-users] failing to respond to cache pressure

2018-08-20 Thread Zhenshi Zhou
et to monitor your MDS in case something goes > > wrong. > > > > Regards, > > Eugen > > > > > > Zitat von Wido den Hollander : > > > >> On 08/13/2018 01:22 PM, Zhenshi Zhou wrote: > >>> Hi, > >>> Recently, the cluster

Re: [ceph-users] failing to respond to cache pressure

2018-08-16 Thread Zhenshi Zhou
Hi, Eugen, I set mds_cache_memory_limit from 1G to 2G, and then to 4G. I still get the warning messages, and the messages would disappear in 1 or 2 minutes. Which version do your kernels use? Zhenshi Zhou 于2018年8月13日周一 下午10:15写道: > Hi Eugen, > The command shows "mds_cache_m

Re: [ceph-users] cephfs kernel client hangs

2018-08-14 Thread Zhenshi Zhou
kernel client Yan, Zheng 于2018年8月14日周二 下午3:13写道: > On Mon, Aug 13, 2018 at 9:55 PM Zhenshi Zhou wrote: > > > > Hi Burkhard, > > I'm sure the user has permission to read and write. Besides, we're not > using EC data pools. > > Now the situation is tha

Re: [ceph-users] failing to respond to cache pressure

2018-08-13 Thread Zhenshi Zhou
t;, "num": 4 } } } # ceph fs status cephfs - 19 clients == +--++---+---+---+---+ | Rank | State |MDS|Activity | dns | inos | +--++---+---+---+---+ | 0 | active | ceph-mon1

Re: [ceph-users] failing to respond to cache pressure

2018-08-13 Thread Zhenshi Zhou
g again (for now). > > Try playing with the cache size to find a setting fitting your needs, > but don't forget to monitor your MDS in case something goes wrong. > > Regards, > Eugen > > > Zitat von Wido den Hollander : > > > On 08/13/2018 01:22 PM, Zhenshi Zhou

Re: [ceph-users] cephfs kernel client hangs

2018-08-13 Thread Zhenshi Zhou
gt; > > On 08/13/2018 03:22 PM, Zhenshi Zhou wrote: > > Hi, > > Finally, I got a running server with files /sys/kernel/debug/ceph/xxx/ > > > > [root@docker27 525c4413-7a08-40ca-9a98-0a6df009025b.client213522]# cat > mdsc > > [root@docker27 525c4413-7a08-40ca-9

Re: [ceph-users] cephfs kernel client hangs

2018-08-13 Thread Zhenshi Zhou
1 PM, Yan, Zheng wrote: > > try 'mount -f', recent kernel should handle 'mount -f' pretty well > > On Wed, Aug 8, 2018 at 10:46 PM Zhenshi Zhou > wrote: > >> Hi, > >> Is there any other way excpet rebooting the server when the client > hangs? &

[ceph-users] failing to respond to cache pressure

2018-08-13 Thread Zhenshi Zhou
Hi, Recently, the cluster runs healthy, but I get warning messages everyday: 2018-08-13 17:39:23.682213 [INF] Cluster is now healthy 2018-08-13 17:39:23.682144 [INF] Health check cleared: MDS_CLIENT_RECALL (was: 6 clients failing to respond to cache pressure) 2018-08-13 17:39:23.052022 [INF] MD

Re: [ceph-users] cephfs kernel client hangs

2018-08-08 Thread Zhenshi Zhou
gia > *Belo Horizonte - Brasil* > *IRC NICK - WebertRLZ* > > > On Wed, Aug 8, 2018 at 11:17 AM Zhenshi Zhou wrote: > >> Hi Webert, >> That command shows the current sessions, whereas the server which I get >> the files(osdc,mdsc,monc) disconnect for a long time.

Re: [ceph-users] cephfs kernel client hangs

2018-08-08 Thread Zhenshi Zhou
the MDS server by issuing `ceph > daemon mds.XX session ls` > > Regards, > > Webert Lima > DevOps Engineer at MAV Tecnologia > *Belo Horizonte - Brasil* > *IRC NICK - WebertRLZ* > > > On Wed, Aug 8, 2018 at 5:08 AM Zhenshi Zhou wrote: > >> Hi, I find a

Re: [ceph-users] cephfs kernel client hangs

2018-08-08 Thread Zhenshi Zhou
it mean? Zhenshi Zhou 于2018年8月8日周三 下午1:58写道: > I restarted the client server so that there's no file in that directory. I > will take care of it if the client hangs next time. > > Thanks > > Yan, Zheng 于2018年8月8日周三 上午11:23写道: > >> On Wed, Aug 8, 2018 at 11:02 AM Z

Re: [ceph-users] cephfs kernel client hangs

2018-08-07 Thread Zhenshi Zhou
I restarted the client server so that there's no file in that directory. I will take care of it if the client hangs next time. Thanks Yan, Zheng 于2018年8月8日周三 上午11:23写道: > On Wed, Aug 8, 2018 at 11:02 AM Zhenshi Zhou wrote: > > > > Hi, > > I check all my ceph serv

Re: [ceph-users] cephfs kernel client hangs

2018-08-07 Thread Zhenshi Zhou
to my question, there must be something else cause the client hang. Zhenshi Zhou 于2018年8月8日周三 上午4:16写道: > Hi, I'm not sure if it just mounts the cephfs without using or doing any > operation within the mounted directory would be affected by flushing cache. > I mounted cephfs on osd serve

Re: [ceph-users] cephfs kernel client hangs

2018-08-07 Thread Zhenshi Zhou
at this > point, but that isn’t etched in stone. > > > > Curious if there is more to share. > > > > Reed > > > > On Aug 7, 2018, at 9:47 AM, Webert de Souza Lima > wrote: > > > > > > Yan, Zheng 于2018年8月7日周二 下午7:51写道: > >> > &

Re: [ceph-users] cephfs kernel client hangs

2018-08-07 Thread Zhenshi Zhou
Hi Yan, thanks for the advice. I will umount the cephfs on osd servers and keep an eye on it.:) Yan, Zheng 于2018年8月7日周二 下午7:51写道: > On Tue, Aug 7, 2018 at 7:15 PM Zhenshi Zhou wrote: > > > > Yes, some osd servers mount cephfs > > > > this can cause memory deadlock.

Re: [ceph-users] cephfs kernel client hangs

2018-08-07 Thread Zhenshi Zhou
Yes, some osd servers mount cephfs Yan, Zheng 于2018年8月7日 周二19:12写道: > did you mount cephfs on the same machines that run ceph-osd? > > On Tue, Aug 7, 2018 at 5:14 PM Zhenshi Zhou wrote: > > > > Hi Burkhard, > > Files located in /sys/kernel/debug/ceph/ are all ne

Re: [ceph-users] cephfs kernel client hangs

2018-08-07 Thread Zhenshi Zhou
Hi Burkhard, Files located in /sys/kernel/debug/ceph/ are all new files generated after I reboot the server. The clients were in blacklist and I manully remove them from the blacklist. But the clients hung still. Thanks Burkhard Linke 于2018年8月7日周二 下午4:54写道: > Hi, > > > you are using the kernel

[ceph-users] cephfs kernel client hangs

2018-08-07 Thread Zhenshi Zhou
Hi, I have a CEPH 12.2.5 cluster running on 4 CentOS 7.3 servers with kernel 4.17.0, Including 3 mons, 16 osds, 2 mds(1active+1backup). I have some cllients mounted cephfs in kernel mode. Client A is using kernel 4.4.145, and others are using kernel 4.12.8. All of them are using ceph client versi