I had the same thing happen too when I built a ceph cluster on a single VM
for testing, I wasn't concerned though because I knew the slow speed was
likely a problem.


On Mon, Oct 15, 2018 at 7:34 AM Kisik Jeong <kisik.je...@csl.skku.edu>
wrote:

> Hello,
>
> I successfully deployed Ceph cluster with 16 OSDs and created CephFS
> before.
> But after rebooting due to mds slow request problem, when creating CephFS,
> Ceph mds goes creating status and never changes.
> Seeing Ceph status, there is no other problem I think. Here is 'ceph -s'
> result:
>
> csl@hpc1:~$ ceph -s
>   cluster:
>     id:     1a32c483-cb2e-4ab3-ac60-02966a8fd327
>     health: HEALTH_OK
>
>   services:
>     mon: 1 daemons, quorum hpc1
>     mgr: hpc1(active)
>     mds: cephfs-1/1/1 up  {0=hpc1=up:creating}
>     osd: 16 osds: 16 up, 16 in
>
>   data:
>     pools:   2 pools, 640 pgs
>     objects: 7 objects, 124B
>     usage:   34.3GiB used, 116TiB / 116TiB avail
>     pgs:     640 active+clean
>
> However, CephFS still works in case of 8 OSDs.
>
> If there is any doubt of this phenomenon, please let me know. Thank you.
>
> PS. I attached my ceph.conf contents:
>
> [global]
> fsid = 1a32c483-cb2e-4ab3-ac60-02966a8fd327
> mon_initial_members = hpc1
> mon_host = 192.168.40.10
> auth_cluster_required = cephx
> auth_service_required = cephx
> auth_client_required = cephx
>
> public_network = 192.168.40.0/24
> cluster_network = 192.168.40.0/24
>
> [osd]
> osd journal size = 1024
> osd max object name len = 256
> osd max object namespace len = 64
> osd mount options f2fs = active_logs=2
>
> [osd.0]
> host = hpc9
> public_addr = 192.168.40.18
> cluster_addr = 192.168.40.18
>
> [osd.1]
> host = hpc10
> public_addr = 192.168.40.19
> cluster_addr = 192.168.40.19
>
> [osd.2]
> host = hpc9
> public_addr = 192.168.40.18
> cluster_addr = 192.168.40.18
>
> [osd.3]
> host = hpc10
> public_addr = 192.168.40.19
> cluster_addr = 192.168.40.19
>
> [osd.4]
> host = hpc9
> public_addr = 192.168.40.18
> cluster_addr = 192.168.40.18
>
> [osd.5]
> host = hpc10
> public_addr = 192.168.40.19
> cluster_addr = 192.168.40.19
>
> [osd.6]
> host = hpc9
> public_addr = 192.168.40.18
> cluster_addr = 192.168.40.18
>
> [osd.7]
> host = hpc10
> public_addr = 192.168.40.19
> cluster_addr = 192.168.40.19
>
> [osd.8]
> host = hpc9
> public_addr = 192.168.40.18
> cluster_addr = 192.168.40.18
>
> [osd.9]
> host = hpc10
> public_addr = 192.168.40.19
> cluster_addr = 192.168.40.19
>
> [osd.10]
> host = hpc9
> public_addr = 192.168.10.18
> cluster_addr = 192.168.40.18
>
> [osd.11]
> host = hpc10
> public_addr = 192.168.10.19
> cluster_addr = 192.168.40.19
>
> [osd.12]
> host = hpc9
> public_addr = 192.168.10.18
> cluster_addr = 192.168.40.18
>
> [osd.13]
> host = hpc10
> public_addr = 192.168.10.19
> cluster_addr = 192.168.40.19
>
> [osd.14]
> host = hpc9
> public_addr = 192.168.10.18
> cluster_addr = 192.168.40.18
>
> [osd.15]
> host = hpc10
> public_addr = 192.168.10.19
> cluster_addr = 192.168.40.19
>
> --
> Kisik Jeong
> Ph.D. Student
> Computer Systems Laboratory
> Sungkyunkwan University
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to