Den 21 jul 2016 8:39 em skrev Scott <[email protected]>: > > CentOS 7 for me as well, using the zfsonlinux.org packages.
Ok, so whatever problems there may be in those packages, at least you have them in common. That can also be a comfort:) /K > > On Thu, Jul 21, 2016 at 1:26 PM David Gossage <[email protected]> > wrote: >> >> On Thu, Jul 21, 2016 at 1:24 PM, Karli Sjöberg <[email protected]> wrote: >>> >>> >>> Den 21 jul 2016 7:54 em skrev David Gossage <[email protected]>: >>> > >>> > On Thu, Jul 21, 2016 at 11:47 AM, Scott <[email protected]> wrote: >>> >> >>> >> Hi David, >>> >> >>> >> My backend storage is ZFS. >>> >> >>> >> I thought about moving from FUSE to NFS mounts for my Gluster volumes to >>> >> help test. But since I use hosted engine this would be a real pain. >>> >> Its difficult to modify the storage domain type/path in the >>> >> hosted-engine.conf. And I don't want to go through the process of >>> >> re-deploying hosted engine. >>> >> >>> > >>> > I found this >>> > >>> > https://bugzilla.redhat.com/show_bug.cgi?id=1347553 >>> > >>> > Not sure if related. >>> > >>> > But I also have zfs backend, another user in gluster mailing list had >>> > issues and used zfs backend although she used proxmox and got it working >>> > by changing disk to writeback cache I think it was. >>> >>> David and Scott, >>> >>> just out of curiousity, what is the OS under ZFS? >> >> Centos 7 >> >>> >>> /K >>> >>> > >>> > I also use hosted engine, but I run my gluster volume for HE actually on >>> > a LVM separate from zfs on xfs and if i recall it did not have the issues >>> > my gluster on zfs did. I'm wondering now if the issue was zfs settings. >>> > >>> > Hopefully should have a test machone up soon I can play around with more. >>> > >>> >> Scott >>> >> >>> >> On Thu, Jul 21, 2016 at 11:36 AM David Gossage >>> >> <[email protected]> wrote: >>> >>> >>> >>> What back end storage do you run gluster on? xfs/zfs/ext4 etc? >>> >>> >>> >>> David Gossage >>> >>> Carousel Checks Inc. | System Administrator >>> >>> Office 708.613.2284 >>> >>> >>> >>> On Thu, Jul 21, 2016 at 8:18 AM, Scott <[email protected]> wrote: >>> >>>> >>> >>>> I get similar problems with oVirt 4.0.1 and hosted engine. After >>> >>>> upgrading all my hosts to Gluster 3.7.13 (client and server), I get >>> >>>> the following: >>> >>>> >>> >>>> $ sudo hosted-engine --set-maintenance --mode=none >>> >>>> Traceback (most recent call last): >>> >>>> File "/usr/lib64/python2.7/runpy.py", line 162, in >>> >>>> _run_module_as_main >>> >>>> "__main__", fname, loader, pkg_name) >>> >>>> File "/usr/lib64/python2.7/runpy.py", line 72, in _run_code >>> >>>> exec code in run_globals >>> >>>> File >>> >>>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_setup/set_maintenance.py", >>> >>>> line 73, in <module> >>> >>>> if not maintenance.set_mode(sys.argv[1]): >>> >>>> File >>> >>>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_setup/set_maintenance.py", >>> >>>> line 61, in set_mode >>> >>>> value=m_global, >>> >>>> File >>> >>>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py", >>> >>>> line 259, in set_maintenance_mode >>> >>>> str(value)) >>> >>>> File >>> >>>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py", >>> >>>> line 204, in set_global_md_flag >>> >>>> all_stats = broker.get_stats_from_storage(service) >>> >>>> File >>> >>>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py", >>> >>>> line 232, in get_stats_from_storage >>> >>>> result = self._checked_communicate(request) >>> >>>> File >>> >>>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py", >>> >>>> line 260, in _checked_communicate >>> >>>> .format(message or response)) >>> >>>> ovirt_hosted_engine_ha.lib.exceptions.RequestError: Request failed: >>> >>>> failed to read metadata: [Errno 1] Operation not permitted >>> >>>> >>> >>>> If I only upgrade one host, then things will continue to work but my >>> >>>> nodes are constantly healing shards. My logs are also flooded with: >>> >>>> >>> >>>> [2016-07-21 13:15:14.137734] W [fuse-bridge.c:2227:fuse_readv_cbk] >>> >>>> 0-glusterfs-fuse: 274714: READ => -1 gfid=4 >>> >>>> 41f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0041d0 (Operation not >>> >>>> permitted) >>> >>>> The message "W [MSGID: 114031] >>> >>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-data-client-0: remote >>> >>>> operation failed [Operation not permitted]" repeated 6 times between >>> >>>> [2016-07-21 13:13:24.134985] and [2016-07-21 13:15:04.132226] >>> >>>> The message "W [MSGID: 114031] >>> >>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-data-client-1: remote >>> >>>> operation failed [Operation not permitted]" repeated 8 times between >>> >>>> [2016-07-21 13:13:34.133116] and [2016-07-21 13:15:14.137178] >>> >>>> The message "W [MSGID: 114031] >>> >>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-data-client-2: remote >>> >>>> operation failed [Operation not permitted]" repeated 7 times between >>> >>>> [2016-07-21 13:13:24.135071] and [2016-07-21 13:15:14.137666] >>> >>>> [2016-07-21 13:15:24.134647] W [MSGID: 114031] >>> >>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-data-client-0: remote >>> >>>> operation failed [Operation not permitted] >>> >>>> [2016-07-21 13:15:24.134764] W [MSGID: 114031] >>> >>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-data-client-2: remote >>> >>>> operation failed [Operation not permitted] >>> >>>> [2016-07-21 13:15:24.134793] W [fuse-bridge.c:2227:fuse_readv_cbk] >>> >>>> 0-glusterfs-fuse: 274741: READ => -1 >>> >>>> gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0038f4 (Operation >>> >>>> not permitted) >>> >>>> [2016-07-21 13:15:34.135413] W [fuse-bridge.c:2227:fuse_readv_cbk] >>> >>>> 0-glusterfs-fuse: 274756: READ => -1 >>> >>>> gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0041d0 (Operation >>> >>>> not permitted) >>> >>>> [2016-07-21 13:15:44.141062] W [fuse-bridge.c:2227:fuse_readv_cbk] >>> >>>> 0-glusterfs-fuse: 274818: READ => -1 >>> >>>> gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0038f4 (Operation >>> >>>> not permitted) >>> >>>> [2016-07-21 13:15:54.133582] W [MSGID: 114031] >>> >>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-data-client-1: remote >>> >>>> operation failed [Operation not permitted] >>> >>>> [2016-07-21 13:15:54.133629] W [fuse-bridge.c:2227:fuse_readv_cbk] >>> >>>> 0-glusterfs-fuse: 274853: READ => -1 >>> >>>> gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0036d8 (Operation >>> >>>> not permitted) >>> >>>> [2016-07-21 13:16:04.133666] W [fuse-bridge.c:2227:fuse_readv_cbk] >>> >>>> 0-glusterfs-fuse: 274879: READ => -1 >>> >>>> gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0041d0 (Operation >>> >>>> not permitted) >>> >>>> [2016-07-21 13:16:14.134954] W [fuse-bridge.c:2227:fuse_readv_cbk] >>> >>>> 0-glusterfs-fuse: 274894: READ => -1 >>> >>>> gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0036d8 (Operation >>> >>>> not permitted) >>> >>>> >>> >>>> Scott >>> >>>> >>> >>>> >>> >>>> On Thu, Jul 21, 2016 at 6:57 AM Frank Rothenstein >>> >>>> <[email protected]> wrote: >>> >>>>> >>> >>>>> Hey Devid, >>> >>>>> >>> >>>>> I have the very same problem on my test-cluster, despite on running >>> >>>>> ovirt 4.0. >>> >>>>> If you access your volumes via NFS all is fine, problem is FUSE. I >>> >>>>> stayed on 3.7.13, but have no solution yet, now I use NFS. >>> >>>>> >>> >>>>> Frank >>> >>>>> >>> >>>>> Am Donnerstag, den 21.07.2016, 04:28 -0500 schrieb David Gossage: >>> >>>>>> >>> >>>>>> Anyone running one of recent 3.6.x lines and gluster using 3.7.13? >>> >>>>>> I am looking to upgrade gluster from 3.7.11->3.7.13 for some bug >>> >>>>>> fixes, but have been told by users on gluster mail list due to some >>> >>>>>> gluster changes I'd need to change the disk parameters to use >>> >>>>>> writeback cache. Something to do with aio support being removed. >>> >>>>>> >>> >>>>>> I believe this could be done with custom parameters? But I believe >>> >>>>>> strage tests are done using dd and would they fail with current >>> >>>>>> settings then? Last upgrade to 3.7.13 I had to rollback to 3.7.11 >>> >>>>>> due to stability isues where gluster storage would go into down >>> >>>>>> state and always show N/A as space available/used. Even if hosts >>> >>>>>> saw storage still and VM's were running on it on all 3 hosts. >>> >>>>>> >>> >>>>>> Saw a lot of messages like these that went away once gluster >>> >>>>>> rollback finished >>> >>>>>> >>> >>>>>> [2016-07-09 15:27:46.935694] I [fuse-bridge.c:4083:fuse_init] >>> >>>>>> 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.22 >>> >>>>>> kernel 7.22 >>> >>>>>> [2016-07-09 15:27:49.555466] W [MSGID: 114031] >>> >>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-GLUSTER1-client-1: >>> >>>>>> remote operation failed [Operation not permitted] >>> >>>>>> [2016-07-09 15:27:49.556574] W [MSGID: 114031] >>> >>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-GLUSTER1-client-0: >>> >>>>>> remote operation failed [Operation not permitted] >>> >>>>>> [2016-07-09 15:27:49.556659] W [fuse-bridge.c:2227:fuse_readv_cbk] >>> >>>>>> 0-glusterfs-fuse: 80: READ => -1 >>> >>>>>> gfid=deb61291-5176-4b81-8315-3f1cf8e3534d fd=0x7f5224002f68 >>> >>>>>> (Operation not permitted) >>> >>>>>> [2016-07-09 15:27:59.612477] W [MSGID: 114031] >>> >>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-GLUSTER1-client-1: >>> >>>>>> remote operation failed [Operation not permitted] >>> >>>>>> [2016-07-09 15:27:59.613700] W [MSGID: 114031] >>> >>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-GLUSTER1-client-0: >>> >>>>>> remote operation failed [Operation not permitted] >>> >>>>>> [2016-07-09 15:27:59.613781] W [fuse-bridge.c:2227:fuse_readv_cbk] >>> >>>>>> 0-glusterfs-fuse: 168: READ => -1 >>> >>>>>> gfid=deb61291-5176-4b81-8315-3f1cf8e3534d fd=0x7f5224002f68 >>> >>>>>> (Operation not permitted) >>> >>>>>> >>> >>>>>> David Gossage >>> >>>>>> Carousel Checks Inc. | System Administrator >>> >>>>>> Office 708.613.2284 >>> >>>>>> >>> >>>>>> _______________________________________________ >>> >>>>>> >>> >>>>>> Users mailing list >>> >>>>>> >>> >>>>>> [email protected] >>> >>>>>> >>> >>>>>> http://lists.ovirt.org/mailman/listinfo/users >>> >>>>>> >>> >>>>> >>> >>>>> >>> >>>>> ________________________________ >>> >>>>> >>> >>>>> >>> >>>>> >>> >>>>> ______________________________________________________________________________ >>> >>>>> BODDEN-KLINIKEN Ribnitz-Damgarten GmbH >>> >>>>> Sandhufe 2 >>> >>>>> 18311 Ribnitz-Damgarten >>> >>>>> >>> >>>>> Telefon: 03821-700-0 >>> >>>>> Fax: 03821-700-240 >>> >>>>> >>> >>>>> E-Mail: [email protected] Internet: >>> >>>>> http://www.bodden-kliniken.de >>> >>>>> >>> >>>>> Sitz: Ribnitz-Damgarten, Amtsgericht: Stralsund, HRB 2919, >>> >>>>> Steuer-Nr.: 079/133/40188 >>> >>>>> Aufsichtsratsvorsitzende: Carmen Schröter, Geschäftsführer: Dr. Falko >>> >>>>> Milski >>> >>>>> >>> >>>>> Der Inhalt dieser E-Mail ist ausschließlich für den bezeichneten >>> >>>>> Adressaten bestimmt. Wenn Sie nicht der vorge- >>> >>>>> sehene Adressat dieser E-Mail oder dessen Vertreter sein sollten, >>> >>>>> beachten Sie bitte, dass jede Form der Veröf- >>> >>>>> fentlichung, Vervielfältigung oder Weitergabe des Inhalts dieser >>> >>>>> E-Mail unzulässig ist. Wir bitten Sie, sofort den >>> >>>>> Absender zu informieren und die E-Mail zu löschen. >>> >>>>> >>> >>>>> >>> >>>>> Bodden-Kliniken Ribnitz-Damgarten GmbH 2016 >>> >>>>> *** Virenfrei durch Kerio Mail Server und Sophos Antivirus *** >>> >>>>> _______________________________________________ >>> >>>>> Users mailing list >>> >>>>> [email protected] >>> >>>>> http://lists.ovirt.org/mailman/listinfo/users >>> > >>> >
_______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

