Glusterfs now in healing mode: Receiver: [root@clei21 ~]# arcstat.py 1 time read miss miss% dmis dm% pmis pm% mmis mm% arcsz c 13:24:49 0 0 0 0 0 0 0 0 0 4.6G 31G 13:24:50 154 80 51 80 51 0 0 80 51 4.6G 31G 13:24:51 179 62 34 62 34 0 0 62 42 4.6G 31G 13:24:52 148 68 45 68 45 0 0 68 45 4.6G 31G 13:24:53 140 64 45 64 45 0 0 64 45 4.6G 31G 13:24:54 124 48 38 48 38 0 0 48 38 4.6G 31G 13:24:55 157 80 50 80 50 0 0 80 50 4.7G 31G 13:24:56 202 68 33 68 33 0 0 68 41 4.7G 31G 13:24:57 127 54 42 54 42 0 0 54 42 4.7G 31G 13:24:58 126 50 39 50 39 0 0 50 39 4.7G 31G 13:24:59 116 40 34 40 34 0 0 40 34 4.7G 31G
Sender [root@clei22 ~]# arcstat.py 1 time read miss miss% dmis dm% pmis pm% mmis mm% arcsz c 13:28:37 8 2 25 2 25 0 0 2 25 468M 31G 13:28:38 1.2K 727 62 727 62 0 0 525 54 469M 31G 13:28:39 815 508 62 508 62 0 0 376 55 469M 31G 13:28:40 994 624 62 624 62 0 0 450 54 469M 31G 13:28:41 783 456 58 456 58 0 0 338 50 470M 31G 13:28:42 916 541 59 541 59 0 0 390 50 470M 31G 13:28:43 768 437 56 437 57 0 0 313 48 471M 31G 13:28:44 877 534 60 534 60 0 0 393 53 470M 31G 13:28:45 957 630 65 630 65 0 0 450 57 470M 31G 13:28:46 819 479 58 479 58 0 0 357 51 471M 31G On Thu, Mar 2, 2017 at 7:18 PM, Juan Pablo <pablo.localh...@gmail.com> wrote: > hey, > what are you using for zfs? get an arc status and show please > > > 2017-03-02 9:57 GMT-03:00 Arman Khalatyan <arm2...@gmail.com>: > >> no, >> ZFS itself is not on top of lvm. only ssd was spitted by lvm for >> slog(10G) and cache (the rest) >> but in any-case the ssd does not help much on glusterfs/ovirt load it >> has almost 100% cache misses....:( (terrible performance compare with nfs) >> >> >> >> >> >> On Thu, Mar 2, 2017 at 1:47 PM, FERNANDO FREDIANI < >> fernando.fredi...@upx.com> wrote: >> >>> Am I understanding correctly, but you have Gluster on the top of ZFS >>> which is on the top of LVM ? If so, why the usage of LVM was necessary ? I >>> have ZFS with any need of LVM. >>> >>> Fernando >>> >>> On 02/03/2017 06:19, Arman Khalatyan wrote: >>> >>> Hi, >>> I use 3 nodes with zfs and glusterfs. >>> Are there any suggestions to optimize it? >>> >>> host zfs config 4TB-HDD+250GB-SSD: >>> [root@clei22 ~]# zpool status >>> pool: zclei22 >>> state: ONLINE >>> scan: scrub repaired 0 in 0h0m with 0 errors on Tue Feb 28 14:16:07 >>> 2017 >>> config: >>> >>> NAME STATE READ WRITE CKSUM >>> zclei22 ONLINE 0 0 0 >>> HGST_HUS724040ALA640_PN2334PBJ4SV6T1 ONLINE 0 0 0 >>> logs >>> lv_slog ONLINE 0 0 0 >>> cache >>> lv_cache ONLINE 0 0 0 >>> >>> errors: No known data errors >>> >>> Name: >>> GluReplica >>> Volume ID: >>> ee686dfe-203a-4caa-a691-26353460cc48 >>> Volume Type: >>> Replicate (Arbiter) >>> Replica Count: >>> 2 + 1 >>> Number of Bricks: >>> 3 >>> Transport Types: >>> TCP, RDMA >>> Maximum no of snapshots: >>> 256 >>> Capacity: >>> 3.51 TiB total, 190.56 GiB used, 3.33 TiB free >>> >>> >>> _______________________________________________ >>> Users mailing >>> listUsers@ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users >>> >>> >>> >>> _______________________________________________ >>> Users mailing list >>> Users@ovirt.org >>> http://lists.ovirt.org/mailman/listinfo/users >>> >>> >> >> _______________________________________________ >> Users mailing list >> Users@ovirt.org >> http://lists.ovirt.org/mailman/listinfo/users >> >> >
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users