On Fri, Sep 28, 2018, 7:16 PM Hetz Ben Hamo <h...@hetz.biz> wrote: > Hi, > > Gobinda, great work! > > One thing though - the device names (sda, sdb etc..) > > On many servers, it's hard to know which disk is which. Lets say I have 10 > spinning disk + 2 SSD's. Which is sda? what about NVME? worse - sometimes > replacing disks replaces the sda to something else. We used to have the > same problem with NICs and now this has been resolved on CentOS/RHEL 7.X > > Could the HCI part - the disk selection part specifically - give more > details? maybe Disk ID or WWN, or anything that can identify a disk? >
/dev/disk/by-id is the right identifier. During installation, it'd be nice if it could show as much data as possible - sdX, /dev/disk/by-id, size and perhaps manufacturer. Y. > Also - SSD caching, most of the time it is recommended to use 2 drives if > possible for good performance. Can a user select X number of drives? > > Thanks > > > On Fri, Sep 28, 2018 at 6:43 PM Gobinda Das <go...@redhat.com> wrote: > >> Hi All, >> Status update on "Hyperconverged Gluster oVirt support" >> >> Features Completed: >> ================ >> >> cockpit-ovirt >> ------------- >> 1- Asymmetric brick configuration.Brick can be configured per host >> basis i.e. If the user wanted to make use of sdb from host1, sdc from >> host2, and sdd from host3. >> 2- Dedupe and Compression integration via VDO support (see >> https://github.com/dm-vdo/kvdo). Gluster bricks are created on vdo >> devices >> 3- LVM cache configuration support (Configure cache by using fast block >> device such as SSD drive to imrove the performance of a larger and slower >> logical volumes) >> 4- Auto addition of 2nd and 3rd hosts in a 3 node setup during >> deployment >> 5- Auto creation of storage domains based on gluster volumes created >> during setup >> 6- Single node deployment support via Cockpit UI. For details on single >> node deployment - >> https://www.ovirt.org/documentation/gluster-hyperconverged/chap-Single_node_hyperconverged/ >> 7- Gluster Management Dashboard (Dashboard will show the nodes in >> cluster,Volumes and bricks. User can expand the cluster and also can create >> new volume in existing cluster nodes ) >> >> oVirt >> ------- >> 1- Reset brick support from UI to allow users to replace a faulty brick >> 2- Create brick from engine now supports configuring an SSD device as >> lvmcache device when bricks are created on spinning disks >> 3- VDO monitoring >> >> GlusterFS >> --------------- >> Enhancements to performance with fuse by 15x >> 1. Cluster after eager lock change for better detection of multiple >> clients >> 2. Changing qemu option aio to "native" instead of "threads". >> >> end-to-end deployment: >> -------------------------------- >> 1- End to end deployment of a Gluster + Ovirt hyperconverged environment >> using ansible roles ( >> https://github.com/gluster/gluster-ansible/tree/master/playbooks ). The >> only pre-requisite is a CentOS node/oVirt node >> >> Future Plan: >> ========== >> cockpit-ovirt: >> >> 1- ansible-roles integration for deployment >> 2- Support for different volume types >> >> vdsm: >> 1- Python3 compatibility of vdsm-gluster >> 2- Native 4K support >> >> -- >> Thanks, >> Gobinda >> _______________________________________________ >> Devel mailing list -- de...@ovirt.org >> To unsubscribe send an email to devel-le...@ovirt.org >> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >> oVirt Code of Conduct: >> https://www.ovirt.org/community/about/community-guidelines/ >> List Archives: >> https://lists.ovirt.org/archives/list/de...@ovirt.org/message/XCFITLLQTODFK6NIRPBTRKKYCWKO6KBP/ >> > _______________________________________________ > Devel mailing list -- de...@ovirt.org > To unsubscribe send an email to devel-le...@ovirt.org > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/de...@ovirt.org/message/WTSJBFP73RTJV6EO4XYZUAHNTOVXYBLS/ >
_______________________________________________ Gluster-devel mailing list Gluster-devel@gluster.org https://lists.gluster.org/mailman/listinfo/gluster-devel