Hello, I am trying to setup a single hyperconverged node. The disks that I will be using for the creation of my engine and all VMs are on a RAID 5. This volume is using hardware raid with an LSI controller (LSI 9361-4i). I am unsure of the appropriate values to use for the Gluster config. Here is the info about my environment.
[root@vmh ~]# mount sysfs on /sys type sysfs (rw,nosuid,nodev,noexec,relatime,seclabel) proc on /proc type proc (rw,nosuid,nodev,noexec,relatime) devtmpfs on /dev type devtmpfs (rw,nosuid,seclabel,size=65713924k,nr_inodes=16428481,mode=755) securityfs on /sys/kernel/security type securityfs (rw,nosuid,nodev,noexec,relatime) tmpfs on /dev/shm type tmpfs (rw,nosuid,nodev,seclabel) devpts on /dev/pts type devpts (rw,nosuid,noexec,relatime,seclabel,gid=5,mode=620,ptmxmode=000) tmpfs on /run type tmpfs (rw,nosuid,nodev,seclabel,mode=755) tmpfs on /sys/fs/cgroup type tmpfs (ro,nosuid,nodev,noexec,seclabel,mode=755) cgroup on /sys/fs/cgroup/systemd type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd) pstore on /sys/fs/pstore type pstore (rw,nosuid,nodev,noexec,relatime) cgroup on /sys/fs/cgroup/freezer type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,freezer) cgroup on /sys/fs/cgroup/pids type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,pids) cgroup on /sys/fs/cgroup/cpuset type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,cpuset) cgroup on /sys/fs/cgroup/cpu,cpuacct type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,cpuacct,cpu) cgroup on /sys/fs/cgroup/blkio type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,blkio) cgroup on /sys/fs/cgroup/hugetlb type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,hugetlb) cgroup on /sys/fs/cgroup/net_cls,net_prio type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,net_prio,net_cls) cgroup on /sys/fs/cgroup/devices type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,devices) cgroup on /sys/fs/cgroup/perf_event type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,perf_event) cgroup on /sys/fs/cgroup/memory type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,memory) configfs on /sys/kernel/config type configfs (rw,relatime) /dev/mapper/onn_vmh-ovirt--node--ng--4.2.5.1--0.20180821.0+1 on / type ext4 (rw,relatime,seclabel,discard,stripe=16,data=ordered) selinuxfs on /sys/fs/selinux type selinuxfs (rw,relatime) systemd-1 on /proc/sys/fs/binfmt_misc type autofs (rw,relatime,fd=22,pgrp=1,timeout=0,minproto=5,maxproto=5,direct,pipe_ino=43386) debugfs on /sys/kernel/debug type debugfs (rw,relatime) mqueue on /dev/mqueue type mqueue (rw,relatime,seclabel) hugetlbfs on /dev/hugepages type hugetlbfs (rw,relatime,seclabel) hugetlbfs on /dev/hugepages1G type hugetlbfs (rw,relatime,seclabel,pagesize=1G) /dev/mapper/onn_vmh-tmp on /tmp type ext4 (rw,relatime,seclabel,discard,stripe=16,data=ordered) /dev/mapper/onn_vmh-home on /home type ext4 (rw,relatime,seclabel,discard,stripe=16,data=ordered) /dev/sdb1 on /boot type ext4 (rw,relatime,seclabel,data=ordered) /dev/mapper/onn_vmh-var on /var type ext4 (rw,relatime,seclabel,discard,stripe=16,data=ordered) /dev/mapper/onn_vmh-var_log on /var/log type ext4 (rw,relatime,seclabel,discard,stripe=16,data=ordered) /dev/mapper/onn_vmh-var_crash on /var/crash type ext4 (rw,relatime,seclabel,discard,stripe=16,data=ordered) /dev/mapper/onn_vmh-var_log_audit on /var/log/audit type ext4 (rw,relatime,seclabel,discard,stripe=16,data=ordered) sunrpc on /var/lib/nfs/rpc_pipefs type rpc_pipefs (rw,relatime) /dev/mapper/3600605b00a2faca222fb4da81ac9bdb1p1 on /data type ext4 (rw,relatime,seclabel,discard,stripe=64,data=ordered) tmpfs on /run/user/0 type tmpfs (rw,nosuid,nodev,relatime,seclabel,size=13149252k,mode=700) [root@vmh ~]# blkid /dev/sda1: UUID="ce130131-c457-46a0-b6de-e50cc89a6da3" TYPE="ext4" PARTUUID="5fff73ae-70d4-4697-a307-5f68a4c00f4c" /dev/sdb1: UUID="47422043-e5d0-4541-86ff-193f61a779b0" TYPE="ext4" /dev/sdb2: UUID="2f2cf71b-b68e-985a-6433-ed1889595df0" UUID_SUB="a3e32abe-109c-b436-7712-6c9d9f1d57c9" LABEL="vmh.cyber-range.lan:pv00" TYPE="linux_raid_member" /dev/sdc1: UUID="2f2cf71b-b68e-985a-6433-ed1889595df0" UUID_SUB="7132acc1-e210-f645-5df2-e1a1eff7f836" LABEL="vmh.cyber-range.lan:pv00" TYPE="linux_raid_member" /dev/md127: UUID="utB9xu-zva6-j5Ci-3E49-uDya-g2AJ-MQu5Rd" TYPE="LVM2_member" /dev/mapper/onn_vmh-ovirt--node--ng--4.2.5.1--0.20180821.0+1: UUID="8656ff1e-d217-4088-b353-6d2b9f602ce3" TYPE="ext4" /dev/mapper/onn_vmh-swap: UUID="57a165d1-116e-4e64-a694-2618ffa3a79e" TYPE="swap" /dev/mapper/3600605b00a2faca222fb4da81ac9bdb1p1: UUID="ce130131-c457-46a0-b6de-e50cc89a6da3" TYPE="ext4" PARTUUID="dac9e1fc-b0d7-43da-b52c-66bb059d8137" /dev/mapper/onn_vmh-root: UUID="7cc65568-d408-43ab-a793-b6c110d7ba98" TYPE="ext4" /dev/mapper/onn_vmh-home: UUID="c7e344d0-f401-4504-b52e-9b5c6023c10e" TYPE="ext4" /dev/mapper/onn_vmh-tmp: UUID="3323e010-0dab-4166-b15d-d739a09b4c03" TYPE="ext4" /dev/mapper/onn_vmh-var: UUID="c20af647-364d-4a11-8776-8e6aac362425" TYPE="ext4" /dev/mapper/onn_vmh-var_log: UUID="eb2e159a-4d50-411d-912c-c8e11f297fea" TYPE="ext4" /dev/mapper/onn_vmh-var_log_audit: UUID="65aa85b9-c053-4244-b3b0-c5f91de8ad59" TYPE="ext4" /dev/mapper/onn_vmh-var_crash: UUID="579bbef5-bab6-49ce-b8ed-ab92afa85ea8" TYPE="ext4" /dev/mapper/3600605b00a2faca222fb4da81ac9bdb1: PTTYPE="gpt" [root@vmh ~]# lsblk NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT sda 8:0 0 8.2T 0 disk ├─sda1 8:1 0 8.2T 0 part └─3600605b00a2faca222fb4da81ac9bdb1 253:5 0 8.2T 0 mpath └─3600605b00a2faca222fb4da81ac9bdb1p1 253:6 0 8.2T 0 part /data sdb 8:16 0 223.6G 0 disk ├─sdb1 8:17 0 1G 0 part /boot └─sdb2 8:18 0 222.6G 0 part └─md127 9:127 0 222.5G 0 raid1 ├─onn_vmh-pool00_tmeta 253:0 0 1G 0 lvm │ └─onn_vmh-pool00-tpool 253:2 0 173.6G 0 lvm │ ├─onn_vmh-ovirt--node--ng--4.2.5.1--0.20180821.0+1 253:3 0 146.6G 0 lvm / │ ├─onn_vmh-pool00 253:7 0 173.6G 0 lvm │ ├─onn_vmh-root 253:8 0 146.6G 0 lvm │ ├─onn_vmh-home 253:9 0 1G 0 lvm /home │ ├─onn_vmh-tmp 253:10 0 1G 0 lvm /tmp │ ├─onn_vmh-var 253:11 0 15G 0 lvm /var │ ├─onn_vmh-var_log 253:12 0 8G 0 lvm /var/log │ ├─onn_vmh-var_log_audit 253:13 0 2G 0 lvm /var/log/audit │ └─onn_vmh-var_crash 253:14 0 10G 0 lvm /var/crash ├─onn_vmh-pool00_tdata 253:1 0 173.6G 0 lvm │ └─onn_vmh-pool00-tpool 253:2 0 173.6G 0 lvm │ ├─onn_vmh-ovirt--node--ng--4.2.5.1--0.20180821.0+1 253:3 0 146.6G 0 lvm / │ ├─onn_vmh-pool00 253:7 0 173.6G 0 lvm │ ├─onn_vmh-root 253:8 0 146.6G 0 lvm │ ├─onn_vmh-home 253:9 0 1G 0 lvm /home │ ├─onn_vmh-tmp 253:10 0 1G 0 lvm /tmp │ ├─onn_vmh-var 253:11 0 15G 0 lvm /var │ ├─onn_vmh-var_log 253:12 0 8G 0 lvm /var/log │ ├─onn_vmh-var_log_audit 253:13 0 2G 0 lvm /var/log/audit │ └─onn_vmh-var_crash 253:14 0 10G 0 lvm /var/crash └─onn_vmh-swap 253:4 0 4G 0 lvm [SWAP] sdc 8:32 0 223.6G 0 disk └─sdc1 8:33 0 222.6G 0 part └─md127 9:127 0 222.5G 0 raid1 ├─onn_vmh-pool00_tmeta 253:0 0 1G 0 lvm │ └─onn_vmh-pool00-tpool 253:2 0 173.6G 0 lvm │ ├─onn_vmh-ovirt--node--ng--4.2.5.1--0.20180821.0+1 253:3 0 146.6G 0 lvm / │ ├─onn_vmh-pool00 253:7 0 173.6G 0 lvm │ ├─onn_vmh-root 253:8 0 146.6G 0 lvm │ ├─onn_vmh-home 253:9 0 1G 0 lvm /home │ ├─onn_vmh-tmp 253:10 0 1G 0 lvm /tmp │ ├─onn_vmh-var 253:11 0 15G 0 lvm /var │ ├─onn_vmh-var_log 253:12 0 8G 0 lvm /var/log │ ├─onn_vmh-var_log_audit 253:13 0 2G 0 lvm /var/log/audit │ └─onn_vmh-var_crash 253:14 0 10G 0 lvm /var/crash ├─onn_vmh-pool00_tdata 253:1 0 173.6G 0 lvm │ └─onn_vmh-pool00-tpool 253:2 0 173.6G 0 lvm │ ├─onn_vmh-ovirt--node--ng--4.2.5.1--0.20180821.0+1 253:3 0 146.6G 0 lvm / │ ├─onn_vmh-pool00 253:7 0 173.6G 0 lvm │ ├─onn_vmh-root 253:8 0 146.6G 0 lvm │ ├─onn_vmh-home 253:9 0 1G 0 lvm /home │ ├─onn_vmh-tmp 253:10 0 1G 0 lvm /tmp │ ├─onn_vmh-var 253:11 0 15G 0 lvm /var │ ├─onn_vmh-var_log 253:12 0 8G 0 lvm /var/log │ ├─onn_vmh-var_log_audit 253:13 0 2G 0 lvm /var/log/audit │ └─onn_vmh-var_crash 253:14 0 10G 0 lvm /var/crash └─onn_vmh-swap 253:4 0 4G 0 lvm [SWAP] [root@vmh ~]# df -h Filesystem Size Used Avail Use% Mounted on /dev/mapper/onn_vmh-ovirt--node--ng--4.2.5.1--0.20180821.0+1 145G 1.9G 135G 2% / devtmpfs 63G 0 63G 0% /dev tmpfs 63G 4.0K 63G 1% /dev/shm tmpfs 63G 19M 63G 1% /run tmpfs 63G 0 63G 0% /sys/fs/cgroup /dev/mapper/onn_vmh-tmp 976M 2.8M 906M 1% /tmp /dev/mapper/onn_vmh-home 976M 2.6M 907M 1% /home /dev/sdb1 976M 204M 706M 23% /boot /dev/mapper/onn_vmh-var 15G 43M 14G 1% /var /dev/mapper/onn_vmh-var_log 7.8G 58M 7.3G 1% /var/log /dev/mapper/onn_vmh-var_crash 9.8G 37M 9.2G 1% /var/crash /dev/mapper/onn_vmh-var_log_audit 2.0G 7.7M 1.8G 1% /var/log/audit /dev/mapper/3600605b00a2faca222fb4da81ac9bdb1p1 8.2T 90M 7.8T 1% /data tmpfs 13G 0 13G 0% /run/user/0 [root@vmh ~]# lspci | grep RAID 3b:00.0 RAID bus controller: LSI Logic / Symbios Logic MegaRAID SAS-3 3108 [Invader] (rev 02) [root@vmh ~]# lsmod | grep mega megaraid_sas 145373 2 [root@vmh ~]# ip a 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host valid_lft forever preferred_lft forever 2: enp96s0f0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP group default qlen 1000 link/ether 0c:c4:7a:f9:b9:88 brd ff:ff:ff:ff:ff:ff inet 172.30.50.3/24 brd 172.30.50.255 scope global noprefixroute dynamic enp96s0f0 valid_lft 47994sec preferred_lft 47994sec inet6 fe80::119b:61e4:f9:1dca/64 scope link noprefixroute valid_lft forever preferred_lft forever 3: enp96s0f1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP group default qlen 1000 link/ether 0c:c4:7a:f9:b9:89 brd ff:ff:ff:ff:ff:ff inet 172.30.50.4/24 brd 172.30.50.255 scope global noprefixroute dynamic enp96s0f1 valid_lft 47994sec preferred_lft 47994sec inet6 fe80::498f:1c33:7de6:1fd/64 scope link noprefixroute valid_lft forever preferred_lft forever 19: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group default qlen 1000 link/ether 2e:78:2c:51:c8:12 brd ff:ff:ff:ff:ff:ff In DNS I have defined .3 address as vmh.cyber-range.lan and .4 is defined as ovirtbe.cyber-range.lan. .3 is the frontend IP and .4 is the backednd IP. I have done the following so far with my Gluster Config file, I'm sure it needs further edits. [hosts] vmh.cyber-range.lan [disktype] jbod #Possible values raid6, raid10, raid5, jbod [diskcount] @NUMBER_OF_DATA_DISKS@ #Ignored in case of jbod [stripesize] 256 #256 in case of jbod [script1] action=execute ignore_script_errors=no file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d @DEVICE@ #[vdo] # Note: Uncomment if dedupe & compression needs to be enabled on device. Needs kmod-vdo module #action=create #names=@VDO_DEVICE_Name@ #devices=@DEVICE@ #logicalsize=@logical_size@T # Note:logicalsize is 10x physical space on disk ##slabsize=32G # Note: used only when the physical size is few TBs #blockmapcachesize=128M #readcache=enabled #readcachesize=20M #emulate512=enabled #writepolicy=auto [pv] action=create devices=@DEVICE@ # Change to @VDO_DEVICE_name@ if using vdo [vg1] action=create vgname=gluster_vg1 pvname=@DEVICE@ # Change to @VDO_DEVICE_name@ if using vdo [lv1] action=create vgname=gluster_vg1 lvname=engine_lv lvtype=thick size=100GB mount=/gluster_bricks/engine [lv2] action=create vgname=gluster_vg1 poolname=lvthinpool lvtype=thinpool poolmetadatasize=7.7TB size=7.7TB #For example: 18000GB, depending on device capacity. Units to be specified. [lv3] action=create lvname=lv_vmdisks poolname=lvthinpool vgname=gluster_vg1 lvtype=thinlv mount=/gluster_bricks/vmstore virtualsize=@SIZE@ # Units to be specified, for instance 5000GB [lv4] action=create lvname=lv_datadisks poolname=lvthinpool vgname=gluster_vg1 lvtype=thinlv mount=/gluster_bricks/data virtualsize=@SIZE@ # Units to be specified, for instance 5000GB #[lv5] #action=setup-cache #ssd=@SSD_DEVICE@ #vgname=gluster_vg1 #poolname=lvthinpool #cache_lv=lvcache #cache_lvsize=5GB # Provide device size ## cachemode=writeback [shell2] action=execute command=vdsm-tool configure --force [script3] action=execute file=/usr/share/gdeploy/scripts/blacklist_all_disks.sh ignore_script_errors=no [selinux] yes [service3] action=restart service=glusterd slice_setup=yes [firewalld] action=add ports=111/tcp,2049/tcp,54321/tcp,5900/tcp,5900-6923/tcp,5666/tcp,16514/tcp,54322/tcp services=glusterfs [script2] action=execute file=/usr/share/gdeploy/scripts/disable-gluster-hooks.sh [shell3] action=execute command=usermod -a -G gluster qemu [volume] action=create volname=engine transport=tcp key=storage.owner-uid,storage.owner-gid,features.shard,performance.low-prio-threads,performance.strict-o-direct,network.remote-dio,network.ping-timeout,user.cifs,nfs.disable,performance.quick-read,performance.read-ahead,performance.io-cache,cluster.eager-lock value=36,36,on,32,on,off,30,off,on,off,off,off,enable brick_dirs=/gluster_bricks/engine/engine ignore_volume_errors=no [volume2] action=create volname=vmstore transport=tcp key=storage.owner-uid,storage.owner-gid,features.shard,performance.low-prio-threads,performance.strict-o-direct,network.remote-dio,network.ping-timeout,user.cifs,nfs.disable,performance.quick-read,performance.read-ahead,performance.io-cache,cluster.eager-lock value=36,36,on,32,on,off,30,off,on,off,off,off,enable brick_dirs=/gluster_bricks/vmstore/vmstore ignore_volume_errors=no [volume3] action=create volname=data transport=tcp key=storage.owner-uid,storage.owner-gid,features.shard,performance.low-prio-threads,performance.strict-o-direct,network.remote-dio,network.ping-timeout,user.cifs,nfs.disable,performance.quick-read,performance.read-ahead,performance.io-cache,cluster.eager-lock value=36,36,on,32,on,off,30,off,on,off,off,off,enable brick_dirs=/gluster_bricks/data/data ignore_volume_errors=no Can anyone kindly provide the correct values and corrections needed? I am not nearly as good with Linux storage as I'd like to be. Please let me know if any further info is required. Many thanks in advance!!!
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/IPSNXA3GHYFU63J5IGQEOVJVPZ42YPCF/