Hi,

Gobinda, great work!

One thing though - the device names (sda, sdb etc..)

On many servers, it's hard to know which disk is which. Lets say I have 10
spinning disk + 2 SSD's. Which is sda? what about NVME? worse - sometimes
replacing disks replaces the sda to something else. We used to have the
same problem with NICs and now this has been resolved on CentOS/RHEL 7.X

Could the HCI part - the disk selection part specifically - give more
details? maybe Disk ID or WWN, or anything that can identify a disk?

Also - SSD caching, most of the time it is recommended to use 2 drives if
possible for good performance. Can a user select X number of drives?

Thanks


On Fri, Sep 28, 2018 at 6:43 PM Gobinda Das <go...@redhat.com> wrote:

> Hi All,
>  Status update on "Hyperconverged Gluster oVirt support"
>
> Features Completed:
> ================
>
>   cockpit-ovirt
>   -------------
>   1- Asymmetric brick configuration.Brick can be configured per host basis
> i.e. If the user wanted to make use of sdb from host1, sdc from host2, and
> sdd from host3.
>   2- Dedupe and Compression integration via VDO support (see
> https://github.com/dm-vdo/kvdo). Gluster bricks are created on vdo devices
>   3- LVM cache configuration support (Configure cache by using fast block
> device such as SSD drive to imrove the performance of a larger and slower
> logical volumes)
>   4- Auto addition of 2nd and 3rd hosts in a 3 node setup during deployment
>   5- Auto creation of storage domains based on gluster volumes created
> during setup
>   6- Single node deployment support via Cockpit UI. For details on single
> node deployment -
> https://www.ovirt.org/documentation/gluster-hyperconverged/chap-Single_node_hyperconverged/
>   7- Gluster Management Dashboard (Dashboard will show the nodes in
> cluster,Volumes and bricks. User can expand the cluster and also can create
> new volume in existing cluster nodes )
>
>   oVirt
>   -------
>   1- Reset brick support from UI to allow users to replace a faulty brick
>   2- Create brick from engine now supports configuring an SSD device as
> lvmcache device when bricks are created on spinning disks
>   3- VDO monitoring
>
>  GlusterFS
> ---------------
>  Enhancements to performance with fuse by 15x
>  1. Cluster after eager lock change for better detection of multiple
> clients
>  2. Changing qemu option aio to "native" instead of "threads".
>
>  end-to-end deployment:
>  --------------------------------
>  1- End to end deployment of a Gluster + Ovirt hyperconverged environment
> using ansible roles (
> https://github.com/gluster/gluster-ansible/tree/master/playbooks ). The
> only pre-requisite is a CentOS node/oVirt node
>
> Future Plan:
> ==========
>  cockpit-ovirt:
>
>   1- ansible-roles integration for deployment
>   2- Support for different volume types
>
>  vdsm:
>   1- Python3 compatibility of vdsm-gluster
>   2- Native 4K support
>
> --
> Thanks,
> Gobinda
> _______________________________________________
> Devel mailing list -- devel@ovirt.org
> To unsubscribe send an email to devel-le...@ovirt.org
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/devel@ovirt.org/message/XCFITLLQTODFK6NIRPBTRKKYCWKO6KBP/
>
_______________________________________________
Devel mailing list -- devel@ovirt.org
To unsubscribe send an email to devel-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/devel@ovirt.org/message/WTSJBFP73RTJV6EO4XYZUAHNTOVXYBLS/

Reply via email to