Re: [Gluster-devel] [Gluster-users] Gluster Community meeting starting in 15 minutes
On Fri, Jul 18, 2014 at 11:15:37AM +0530, Vijay Bellur wrote: On 07/18/2014 08:54 AM, Pranith Kumar Karampuri wrote: On 07/18/2014 12:08 AM, Vijay Bellur wrote: On 07/17/2014 01:43 AM, Niels de Vos wrote: Meeting summary --- * LINK:https://public.pad.fsfe.org/p/gluster-community-meetings (ndevos, 15:02:45) * roll call (ndevos, 15:03:28) * Action items from last meeting (ndevos, 15:05:31) * LINK:http://review.gluster.org/#/c/8240/(jdarcy, 15:15:05) * glusterfs-3.4 (ndevos, 15:20:13) * glusterfs-3.5 (ndevos, 15:24:26) * LINK: https://bugzilla.redhat.com/showdependencytree.cgi?hide_resolved=1id=glusterfs-3.5.2 (ndevos, 15:25:57) * glusterfs-3.6 (ndevos, 15:31:13) * ACTION: hagarth to branch release-3.6 tomorrow (17th of July) and start the qa-release process (ndevos, 15:33:21) * ACTION: pranithk discuss with vijay about community test days for 3.6 and its potential conflict with a 3.5.2 beta (ndevos, 15:39:23) * ACTION: pranithk propose a date for the community test-days of 3.6, and whatnot (ndevos, 15:40:04) Had a discussion with Pranith and we felt that 3.5.2 beta is of more importance than 3.6 community test days. Hence we are moving the first test day for 3.6 to 2nd August. I have updated the Planning36 page to reflect this. Vijay, Niels wanted to release first beta of 3.5.2 as well on 1st August :-( Can we target 3.5.2 earlier as release-3.5 has received few good bug fixes after 3.5.1 was done? Niels: Are there issues blocking us from doing a 3.5.2 beta earlier? I might be able to do a 3.5.2 beta this weekend or on Monday. After that, I can not guarantee my availability due to a hospital visit and some recovery period. Current status of the glusterfs-3.5.2 blocker (some bugs might need an update of the status): #1081016 ASSIGNED - kkeit...@redhat.com - glusterd needs xfsprogs and e2fsprogs packages #1092037 POST - lmoha...@redhat.com - Issues reported by Cppcheck static analysis tool #1093594 POST - pguru...@redhat.com - Glfs_fini() not freeing the resources #1099270 ASSIGNED - rgowd...@redhat.com - Gluster 3.5.0 NFS server crashes under load #1100050 MODIFIED - kdhan...@redhat.com - Can't write to quota enable folder #1105891 MODIFIED - khire...@redhat.com - features/gfid-access: stat on .gfid virtual directory return EINVAL #454 MODIFIED - ravishan...@redhat.com - creating symlinks generates errors on stripe volume #1112348 MODIFIED - pkara...@redhat.com - [AFR] I/O fails when one of the replica nodes go down #1112980 MODIFIED - nde...@redhat.com - NFS subdir authentication doesn't correctly handle multi-(homed,protocol,etc) network addresses #1113007 MODIFIED - vum...@redhat.com - nfs-utils should be installed as dependency while installing glusterfs-server #1113050 ASSIGNED - kdhan...@redhat.com - Transient failures immediately after add-brick to a mounted volume #1113403 MODIFIED - pkara...@redhat.com - Excessive logging in quotad.log of the kind 'null client' #1113749 MODIFIED - kkeit...@redhat.com - client_t clienttable cliententries are never expanded when all entries are used #1113894 MODIFIED - pkara...@redhat.com - AFR : self-heal of few files not happening when a AWS EC2 Instance is back online after a restart #1113959 MODIFIED - kkeit...@redhat.com - Spec %post server does not wait for the old glusterd to exit #1114501 MODIFIED - pkara...@redhat.com - Dist-geo-rep : deletion of files on master, geo-rep fails to propagate to slaves. #1115369 MODIFIED - nde...@redhat.com - [3.5.2] Allow the usage of the wildcard character '*' to the options nfs.rpc-auth-allow and nfs.rpc-auth-reject #1115950 POST - ravishan...@redhat.com - glfsheal: Improve the way in which we check the presence of replica volumes #1116672 MODIFIED - pkara...@redhat.com - Resource cleanup doesn't happen for clients on servers after disconnect #1116997 MODIFIED - nde...@redhat.com - [3.5.2] mounting a volume over NFS (TCP) with MOUNT over UDP fails #1117241 NEW- kaus...@redhat.com - backport 'gluster volume status --xml' issues #1117888 NEW- rgowd...@redhat.com - Problem when enabling quota : Could not start quota auxiliary mount #1119827 NEW- gluster-b...@redhat.com - Brick goes offline unexpectedly #1120151 POST - gluster-b...@redhat.com - Glustershd memory usage too high #1120245 ASSIGNED - pkara...@redhat.com - Glustershd memory usage too high If the bugs in MODIFIED or POST are sufficient for a 3.5.2 beta, there should be no problem in getting it out soon. Thanks, Niels ___ Gluster-devel mailing list Gluster-devel@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] Inspiration for improving our contributor documentation
On Fri, Jul 18, 2014 at 11:39 AM, Pranith Kumar Karampuri pkara...@redhat.com wrote: On 07/18/2014 11:22 AM, Kaushal M wrote: On Fri, Jul 18, 2014 at 11:11 AM, Pranith Kumar Karampuri pkara...@redhat.com wrote: On 07/17/2014 07:25 PM, Kaushal M wrote: I came across mediawiki's developer documentation and guides when browsing. These docs felt really good to me, and easy to approach. I feel that we should take inspiration from them and start enhancing our docs. (Outright copying with modifications as necessary, could work too. But that just doesn't feel right) Any volunteers? (I'll start as soon as I finish with the developer documentation for data structures for the components I volunteered earlier) ~kaushal [0] - https://www.mediawiki.org/wiki/Developer_hub I love the idea but not sure about the implementation. i.e. considering we already started with .md pages, why not have same kind of pages as .md files in /doc of gluster? We can modify the README in our project so that people can browse all the details in github? Please let me know your thoughts. These kinds of docs need to indexable and searchable by search engines. Only then will they be useful. I don't think markdown files in the source would be good place for these. Still confused :-(. What exactly do you have in mind about things that need to be added to the Developer_hub on the wiki? So lets say we complete writing all this documentation in .md. Can we point to already added documentation on the github repo or do we have to write them again on the wiki? By the way this doubt is only about developer documentation. I completely agree about the rest of the pages you gave below. I was mainly talking about the documents under 'Code, development and style' section in the wikimedia developer hub and our current documentation in those areas. For the other developer documents (arch, api, etc.) we can continue with the current 'markdown in git repo' method. We could add links from the wiki to these. Also there are tools which can serve markdown files and we could probably setup such a server as well. If you are concerned about writing the documentation in 2 different formats (wiki and markdown), we could consolidate to just 1. There are extensions to mediawiki, which allows writing entries in markdown. Pranith The other docs related to source/code documentation can be provided in the source as we are attempting to provide now. These need to be directly accessible for devs when developing, so having them in the git repo is good. Pranith [1] - https://www.mediawiki.org/wiki/Category:New_contributors [2] - https://www.mediawiki.org/wiki/Gerrit/Code_review [3] - https://www.mediawiki.org/wiki/Gerrit [4] - https://www.mediawiki.org/wiki/Gerrit/Tutorial [5] - https://www.mediawiki.org/wiki/Gerrit/Getting_started [6] - https://www.mediawiki.org/wiki/Gerrit/Advanced_usage ... and lots more. ___ Gluster-devel mailing list Gluster-devel@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-devel ___ Gluster-devel mailing list Gluster-devel@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] spurious regression failures again!
Hi, Created a bug against the same. Please use it to submit if required. https://bugzilla.redhat.com/show_bug.cgi?id=1121014 Thanks Varun Shastry On Tuesday 15 July 2014 09:34 PM, Pranith Kumar Karampuri wrote: On 07/15/2014 09:24 PM, Joseph Fernandes wrote: Hi Pranith, Could you please share the link of the console output of the failures. Added them inline. Thanks for reminding :-) Pranith Regards, Joe - Original Message - From: Pranith Kumar Karampuri pkara...@redhat.com To: Gluster Devel gluster-devel@gluster.org, Varun Shastry vshas...@redhat.com Sent: Tuesday, July 15, 2014 8:52:44 PM Subject: [Gluster-devel] spurious regression failures again! hi, We have 4 tests failing once in a while causing problems: 1) tests/bugs/bug-1087198.t - Author: Varun http://build.gluster.org/job/rackspace-regression-2GB-triggered/379/consoleFull 2) tests/basic/mgmt_v3-locks.t - Author: Avra http://build.gluster.org/job/rackspace-regression-2GB-triggered/375/consoleFull 3) tests/basic/fops-sanity.t - Author: Pranith http://build.gluster.org/job/rackspace-regression-2GB-triggered/383/consoleFull Please take a look at them and post updates. Pranith ___ Gluster-devel mailing list Gluster-devel@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-devel ___ Gluster-devel mailing list Gluster-devel@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] [ovirt-users] Can we debug some truths/myths/facts about hosted-engine and gluster?
[Adding gluster-devel] On 07/18/2014 05:20 PM, Andrew Lau wrote: Hi all, As most of you have got hints from previous messages, hosted engine won't work on gluster . A quote from BZ1097639 Using hosted engine with Gluster backed storage is currently something we really warn against. I think this bug should be closed or re-targeted at documentation, because there is nothing we can do here. Hosted engine assumes that all writes are atomic and (immediately) available for all hosts in the cluster. Gluster violates those assumptions. I tried going through BZ1097639 but could not find much detail with respect to gluster there. A few questions around the problem: 1. Can somebody please explain in detail the scenario that causes the problem? 2. Is hosted engine performing synchronous writes to ensure that writes are durable? Also, if there is any documentation that details the hosted engine architecture that would help in enhancing our understanding of its interactions with gluster. Now my question, does this theory prevent a scenario of perhaps something like a gluster replicated volume being mounted as a glusterfs filesystem and then re-exported as the native kernel NFS share for the hosted-engine to consume? It could then be possible to chuck ctdb in there to provide a last resort failover solution. I have tried myself and suggested it to two people who are running a similar setup. Now using the native kernel NFS server for hosted-engine and they haven't reported as many issues. Curious, could anyone validate my theory on this? If we obtain more details on the use case and obtain gluster logs from the failed scenarios, we should be able to understand the problem better. That could be the first step in validating your theory or evolving further recommendations :). Thanks, Vijay ___ Gluster-devel mailing list Gluster-devel@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] [ovirt-users] Can we debug some truths/myths/facts about hosted-engine and gluster?
On Fri, Jul 18, 2014 at 10:06 PM, Vijay Bellur vbel...@redhat.com wrote: [Adding gluster-devel] On 07/18/2014 05:20 PM, Andrew Lau wrote: Hi all, As most of you have got hints from previous messages, hosted engine won't work on gluster . A quote from BZ1097639 Using hosted engine with Gluster backed storage is currently something we really warn against. I think this bug should be closed or re-targeted at documentation, because there is nothing we can do here. Hosted engine assumes that all writes are atomic and (immediately) available for all hosts in the cluster. Gluster violates those assumptions. I tried going through BZ1097639 but could not find much detail with respect to gluster there. A few questions around the problem: 1. Can somebody please explain in detail the scenario that causes the problem? 2. Is hosted engine performing synchronous writes to ensure that writes are durable? Also, if there is any documentation that details the hosted engine architecture that would help in enhancing our understanding of its interactions with gluster. Now my question, does this theory prevent a scenario of perhaps something like a gluster replicated volume being mounted as a glusterfs filesystem and then re-exported as the native kernel NFS share for the hosted-engine to consume? It could then be possible to chuck ctdb in there to provide a last resort failover solution. I have tried myself and suggested it to two people who are running a similar setup. Now using the native kernel NFS server for hosted-engine and they haven't reported as many issues. Curious, could anyone validate my theory on this? If we obtain more details on the use case and obtain gluster logs from the failed scenarios, we should be able to understand the problem better. That could be the first step in validating your theory or evolving further recommendations :). I'm not sure how useful this is, but Jiri Moskovcak tracked this down in an off list message. Message Quote: == We were able to track it down to this (thanks Andrew for providing the testing setup): -b686-4363-bb7e-dba99e5789b6/ha_agent service_type=hosted-engine' Traceback (most recent call last): File /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/broker/listener.py, line 165, in handle response = success + self._dispatch(data) File /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/broker/listener.py, line 261, in _dispatch .get_all_stats_for_service_type(**options) File /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/broker/storage_broker.py, line 41, in get_all_stats_for_service_type d = self.get_raw_stats_for_service_type(storage_dir, service_type) File /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/broker/storage_broker.py, line 74, in get_raw_stats_for_service_type f = os.open(path, direct_flag | os.O_RDONLY) OSError: [Errno 116] Stale file handle: '/rhev/data-center/mnt/localho st:_mnt_hosted-engine/c898fd2a-b686-4363-bb7e-dba99e5789b6/ha_agent/hosted- engine.metadata' It's definitely connected to the storage which leads us to the gluster, I'm not very familiar with the gluster so I need to check this with our gluster gurus. == Thanks, Vijay ___ Gluster-devel mailing list Gluster-devel@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] release-3.6 branch created
On 2014-07-17 20:22, Vijay Bellur wrote: Hi All, A new branch, 'release-3.6', has been branched from this commit in master: commit 950f9d8abe714708ca62b86f304e7417127e1132 Author: Jeff Darcy jda...@redhat.com Date: Tue Jul 8 21:56:04 2014 -0400 dht: fix rename race You can checkout this branch through: $git checkout -b release-3.6 origin/release-3.6 rfc.sh is being updated to send patches to the appropriate branch. The plan is to have all 3.6.x releases happen off this branch. If you need any fix to be part of a 3.4.x release, please send out a backport of the same from master to release-3.4 after it has been accepted in master. More notes on backporting are available at [1]. Shouldn't the root of this branch get a tag to avoid this weirdness: # git checkout -b release-3.6 origin/release-3.6 Branch release-3.6 set up to track remote branch release-3.6 from origin. Switched to a new branch 'release-3.6' # git describe v3.5qa2-762-g950f9d8 or have I missed some git magic? /Anders -- Anders Blomdell Email: anders.blomd...@control.lth.se Department of Automatic Control Lund University Phone:+46 46 222 4625 P.O. Box 118 Fax: +46 46 138118 SE-221 00 Lund, Sweden ___ Gluster-devel mailing list Gluster-devel@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] Duplicate entries and other weirdness in a 3*4 volume
On 07/18/2014 07:57 PM, Anders Blomdell wrote: During testing of a 3*4 gluster (from master as of yesterday), I encountered two major weirdnesses: 1. A 'rm -rf some_dir' needed several invocations to finish, each time reporting a number of lines like these: rm: cannot remove ‘a/b/c/d/e/f’: Directory not empty 2. After having successfully deleted all files from the volume, i have a single directory that is duplicated in gluster-fuse, like this: # ls -l /mnt/gluster total 24 drwxr-xr-x 2 root root 12288 18 jul 16.17 work2/ drwxr-xr-x 2 root root 12288 18 jul 16.17 work2/ any idea on how to debug this issue? /Anders Anders, Check Gluster log files present in /var/log/glusterfs. Specifically glusterd logfile i.e. /var/log/glusterfs/etc-glusterfs-glusterd.vol.log. You can also start glusterd with debug mode i.e. $glusterd -L DEBUG and check the log files for more information. Thanks, Lala ___ Gluster-devel mailing list Gluster-devel@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] Duplicate entries and other weirdness in a 3*4 volume
On 07/18/2014 07:57 PM, Anders Blomdell wrote: During testing of a 3*4 gluster (from master as of yesterday), I encountered two major weirdnesses: 1. A 'rm -rf some_dir' needed several invocations to finish, each time reporting a number of lines like these: rm: cannot remove ‘a/b/c/d/e/f’: Directory not empty 2. After having successfully deleted all files from the volume, i have a single directory that is duplicated in gluster-fuse, like this: # ls -l /mnt/gluster total 24 drwxr-xr-x 2 root root 12288 18 jul 16.17 work2/ drwxr-xr-x 2 root root 12288 18 jul 16.17 work2/ any idea on how to debug this issue? What are the steps to recreate? We need to first find what lead to this. Then probably which xlator leads to this. Pranith /Anders ___ Gluster-devel mailing list Gluster-devel@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-devel