Re: [Gluster-devel] [Gluster-users] Gluster Community meeting starting in 15 minutes

2014-07-18 Thread Niels de Vos
On Fri, Jul 18, 2014 at 11:15:37AM +0530, Vijay Bellur wrote:
 On 07/18/2014 08:54 AM, Pranith Kumar Karampuri wrote:
 
 On 07/18/2014 12:08 AM, Vijay Bellur wrote:
 On 07/17/2014 01:43 AM, Niels de Vos wrote:
 Meeting summary
 ---
 * LINK:https://public.pad.fsfe.org/p/gluster-community-meetings
(ndevos, 15:02:45)
 * roll call  (ndevos, 15:03:28)
 
 * Action items from last meeting  (ndevos, 15:05:31)
* LINK:http://review.gluster.org/#/c/8240/(jdarcy, 15:15:05)
 
 * glusterfs-3.4  (ndevos, 15:20:13)
 
 * glusterfs-3.5  (ndevos, 15:24:26)
* LINK:
 https://bugzilla.redhat.com/showdependencytree.cgi?hide_resolved=1id=glusterfs-3.5.2
 
  (ndevos, 15:25:57)
 
 * glusterfs-3.6  (ndevos, 15:31:13)
* ACTION: hagarth to branch release-3.6 tomorrow (17th of July) and
  start the qa-release process  (ndevos, 15:33:21)
* ACTION: pranithk discuss with vijay about community test days for
  3.6 and its potential conflict with a 3.5.2 beta  (ndevos,
 15:39:23)
* ACTION: pranithk propose a date for the community test-days of 3.6,
  and whatnot  (ndevos, 15:40:04)
 
 Had a discussion with Pranith and we felt that 3.5.2 beta is of more
 importance than 3.6 community test days. Hence we are moving the first
 test day for 3.6 to 2nd August. I have updated the Planning36 page to
 reflect this.
 Vijay,
   Niels wanted to release first beta of 3.5.2 as well on 1st August :-(
 
 
 Can we target 3.5.2 earlier as release-3.5 has received few good bug
 fixes after 3.5.1 was done?
 
 Niels: Are there issues blocking us from doing a 3.5.2 beta earlier?

I might be able to do a 3.5.2 beta this weekend or on Monday. After 
that, I can not guarantee my availability due to a hospital visit and 
some recovery period.

Current status of the glusterfs-3.5.2 blocker (some bugs might need an 
update of the status):

#1081016 ASSIGNED   - kkeit...@redhat.com - glusterd needs xfsprogs and 
e2fsprogs packages
#1092037 POST   - lmoha...@redhat.com - Issues reported by Cppcheck static 
analysis tool
#1093594 POST   - pguru...@redhat.com - Glfs_fini() not freeing the 
resources
#1099270 ASSIGNED   - rgowd...@redhat.com - Gluster 3.5.0 NFS server crashes 
under load
#1100050 MODIFIED   - kdhan...@redhat.com - Can't write to quota enable folder
#1105891 MODIFIED   - khire...@redhat.com - features/gfid-access: stat on .gfid 
virtual directory return EINVAL
#454 MODIFIED   - ravishan...@redhat.com - creating symlinks generates 
errors on stripe volume
#1112348 MODIFIED   - pkara...@redhat.com - [AFR] I/O fails when one of the 
replica nodes go down
#1112980 MODIFIED   - nde...@redhat.com - NFS subdir authentication doesn't 
correctly handle multi-(homed,protocol,etc) network addresses
#1113007 MODIFIED   - vum...@redhat.com - nfs-utils should be installed as 
dependency while installing glusterfs-server
#1113050 ASSIGNED   - kdhan...@redhat.com - Transient failures immediately 
after add-brick to a mounted volume
#1113403 MODIFIED   - pkara...@redhat.com - Excessive logging in quotad.log of 
the kind 'null client'
#1113749 MODIFIED   - kkeit...@redhat.com - client_t clienttable cliententries 
are never expanded when all entries are used
#1113894 MODIFIED   - pkara...@redhat.com - AFR : self-heal of few files not 
happening when a AWS EC2 Instance is back online after a restart
#1113959 MODIFIED   - kkeit...@redhat.com - Spec %post server does not wait for 
the old glusterd to exit
#1114501 MODIFIED   - pkara...@redhat.com - Dist-geo-rep : deletion of files on 
master,  geo-rep fails to propagate to slaves.
#1115369 MODIFIED   - nde...@redhat.com - [3.5.2] Allow the usage of the 
wildcard character '*' to the options nfs.rpc-auth-allow and 
nfs.rpc-auth-reject
#1115950 POST   - ravishan...@redhat.com - glfsheal: Improve the way in 
which we check the presence of replica volumes
#1116672 MODIFIED   - pkara...@redhat.com - Resource cleanup doesn't happen for 
clients on servers after disconnect
#1116997 MODIFIED   - nde...@redhat.com - [3.5.2] mounting a volume over NFS 
(TCP) with MOUNT over UDP fails
#1117241 NEW- kaus...@redhat.com - backport 'gluster volume status 
--xml' issues
#1117888 NEW- rgowd...@redhat.com - Problem when enabling quota : Could 
not start quota auxiliary mount
#1119827 NEW- gluster-b...@redhat.com - Brick goes offline unexpectedly
#1120151 POST   - gluster-b...@redhat.com - Glustershd memory usage too high
#1120245 ASSIGNED   - pkara...@redhat.com - Glustershd memory usage too high

If the bugs in MODIFIED or POST are sufficient for a 3.5.2 beta, there 
should be no problem in getting it out soon.

Thanks,
Niels
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] Inspiration for improving our contributor documentation

2014-07-18 Thread Kaushal M
On Fri, Jul 18, 2014 at 11:39 AM, Pranith Kumar Karampuri
pkara...@redhat.com wrote:

 On 07/18/2014 11:22 AM, Kaushal M wrote:

 On Fri, Jul 18, 2014 at 11:11 AM, Pranith Kumar Karampuri
 pkara...@redhat.com wrote:

 On 07/17/2014 07:25 PM, Kaushal M wrote:

 I came across mediawiki's developer documentation and guides when
 browsing. These docs felt really good to me, and easy to approach.
 I feel that we should take inspiration from them and start enhancing
 our docs. (Outright copying with modifications as necessary, could
 work too. But that just doesn't feel right)

 Any volunteers?
 (I'll start as soon as I finish with the developer documentation for
 data structures for the components I volunteered earlier)

 ~kaushal

 [0] - https://www.mediawiki.org/wiki/Developer_hub

 I love the idea but not sure about the implementation. i.e. considering
 we
 already started with .md pages, why not have same kind of pages as .md
 files
 in /doc of gluster? We can modify the README in our project so that
 people
 can browse all the details in github? Please let me know your thoughts.

 These kinds of docs need to indexable and searchable by search
 engines. Only then will they be useful. I don't think markdown files
 in the source would be good place for these.

 Still confused :-(. What exactly do you have in mind about things that need
 to be added to the Developer_hub on the wiki? So lets say we complete
 writing all this documentation in .md. Can we point to already added
 documentation on the github repo or do we have to write them again on the
 wiki? By the way this doubt is only about developer documentation. I
 completely agree about the rest of the pages you gave below.

I was mainly talking about the documents under 'Code, development and
style' section in the wikimedia developer hub and our current
documentation in those areas.
For the other developer documents (arch, api, etc.) we can continue
with the current 'markdown in git repo' method. We could add links
from the wiki to these. Also there are tools which can serve markdown
files and we could probably setup such a server as well.

If you are concerned about writing the documentation in 2 different
formats (wiki and markdown), we could consolidate to just 1. There are
extensions to mediawiki, which allows writing entries in markdown.


 Pranith

 The other docs related to source/code documentation can be provided in
 the source as we are attempting to provide now. These need to be
 directly accessible for devs when developing, so having them in the
 git repo is good.

 Pranith

 [1] - https://www.mediawiki.org/wiki/Category:New_contributors
 [2] - https://www.mediawiki.org/wiki/Gerrit/Code_review
 [3] - https://www.mediawiki.org/wiki/Gerrit
 [4] - https://www.mediawiki.org/wiki/Gerrit/Tutorial
 [5] - https://www.mediawiki.org/wiki/Gerrit/Getting_started
 [6] - https://www.mediawiki.org/wiki/Gerrit/Advanced_usage
 ... and lots more.
 ___
 Gluster-devel mailing list
 Gluster-devel@gluster.org
 http://supercolony.gluster.org/mailman/listinfo/gluster-devel



___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] spurious regression failures again!

2014-07-18 Thread Varun Shastry

Hi,

Created a bug against the same. Please use it to submit if required.
https://bugzilla.redhat.com/show_bug.cgi?id=1121014

Thanks
Varun Shastry


On Tuesday 15 July 2014 09:34 PM, Pranith Kumar Karampuri wrote:


On 07/15/2014 09:24 PM, Joseph Fernandes wrote:

Hi Pranith,

Could you please share the link of the console output of the failures.

Added them inline. Thanks for reminding :-)

Pranith


Regards,
Joe

- Original Message -
From: Pranith Kumar Karampuri pkara...@redhat.com
To: Gluster Devel gluster-devel@gluster.org, Varun Shastry 
vshas...@redhat.com

Sent: Tuesday, July 15, 2014 8:52:44 PM
Subject: [Gluster-devel] spurious regression failures again!

hi,
  We have 4 tests failing once in a while causing problems:
1) tests/bugs/bug-1087198.t - Author: Varun
http://build.gluster.org/job/rackspace-regression-2GB-triggered/379/consoleFull 


2) tests/basic/mgmt_v3-locks.t - Author: Avra
http://build.gluster.org/job/rackspace-regression-2GB-triggered/375/consoleFull 


3) tests/basic/fops-sanity.t - Author: Pranith
http://build.gluster.org/job/rackspace-regression-2GB-triggered/383/consoleFull 



Please take a look at them and post updates.

Pranith
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel




___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] [ovirt-users] Can we debug some truths/myths/facts about hosted-engine and gluster?

2014-07-18 Thread Vijay Bellur

[Adding gluster-devel]

On 07/18/2014 05:20 PM, Andrew Lau wrote:

Hi all,

As most of you have got hints from previous messages, hosted engine
won't work on gluster . A quote from BZ1097639

Using hosted engine with Gluster backed storage is currently something
we really warn against.


I think this bug should be closed or re-targeted at documentation, because 
there is nothing we can do here. Hosted engine assumes that all writes are 
atomic and (immediately) available for all hosts in the cluster. Gluster 
violates those assumptions.
​
I tried going through BZ1097639 but could not find much detail with 
respect to gluster there.


A few questions around the problem:

1. Can somebody please explain in detail the scenario that causes the 
problem?


2. Is hosted engine performing synchronous writes to ensure that writes 
are durable?


Also, if there is any documentation that details the hosted engine 
architecture that would help in enhancing our understanding of its 
interactions with gluster.



​

Now my question, does this theory prevent a scenario of perhaps
something like a gluster replicated volume being mounted as a glusterfs
filesystem and then re-exported as the native kernel NFS share for the
hosted-engine to consume? It could then be possible to chuck ctdb in
there to provide a last resort failover solution. I have tried myself
and suggested it to two people who are running a similar setup. Now
using the native kernel NFS server for hosted-engine and they haven't
reported as many issues. Curious, could anyone validate my theory on this?



If we obtain more details on the use case and obtain gluster logs from 
the failed scenarios, we should be able to understand the problem 
better. That could be the first step in validating your theory or 
evolving further recommendations :).


Thanks,
Vijay
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] [ovirt-users] Can we debug some truths/myths/facts about hosted-engine and gluster?

2014-07-18 Thread Andrew Lau
​​

On Fri, Jul 18, 2014 at 10:06 PM, Vijay Bellur vbel...@redhat.com wrote:

 [Adding gluster-devel]


 On 07/18/2014 05:20 PM, Andrew Lau wrote:

 Hi all,

 As most of you have got hints from previous messages, hosted engine
 won't work on gluster . A quote from BZ1097639

 Using hosted engine with Gluster backed storage is currently something
 we really warn against.


 I think this bug should be closed or re-targeted at documentation,
 because there is nothing we can do here. Hosted engine assumes that all
 writes are atomic and (immediately) available for all hosts in the cluster.
 Gluster violates those assumptions.
 ​

 I tried going through BZ1097639 but could not find much detail with
 respect to gluster there.

 A few questions around the problem:

 1. Can somebody please explain in detail the scenario that causes the
 problem?

 2. Is hosted engine performing synchronous writes to ensure that writes
 are durable?

 Also, if there is any documentation that details the hosted engine
 architecture that would help in enhancing our understanding of its
 interactions with gluster.


  ​

 Now my question, does this theory prevent a scenario of perhaps
 something like a gluster replicated volume being mounted as a glusterfs
 filesystem and then re-exported as the native kernel NFS share for the
 hosted-engine to consume? It could then be possible to chuck ctdb in
 there to provide a last resort failover solution. I have tried myself
 and suggested it to two people who are running a similar setup. Now
 using the native kernel NFS server for hosted-engine and they haven't
 reported as many issues. Curious, could anyone validate my theory on this?


 If we obtain more details on the use case and obtain gluster logs from the
 failed scenarios, we should be able to understand the problem better. That
 could be the first step in validating your theory or evolving further
 recommendations :).


​I'm not sure how useful this is, but ​Jiri Moskovcak tracked this down in
an off list message.

​Message Quote:​

​==​

​We were able to track it down to this (thanks Andrew for providing the
testing setup):

-b686-4363-bb7e-dba99e5789b6/ha_agent service_type=hosted-engine'
Traceback (most recent call last):
  File 
/usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/broker/listener.py,
line 165, in handle
response = success  + self._dispatch(data)
  File 
/usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/broker/listener.py,
line 261, in _dispatch
.get_all_stats_for_service_type(**options)
  File 
/usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/broker/storage_broker.py,
line 41, in get_all_stats_for_service_type
d = self.get_raw_stats_for_service_type(storage_dir, service_type)
  File 
/usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/broker/storage_broker.py,
line 74, in get_raw_stats_for_service_type
f = os.open(path, direct_flag | os.O_RDONLY)
OSError: [Errno 116] Stale file handle: '/rhev/data-center/mnt/localho
st:_mnt_hosted-engine/c898fd2a-b686-4363-bb7e-dba99e5789b6/ha_agent/hosted-
engine.metadata'

It's definitely connected to the storage which leads us to the gluster, I'm
not very familiar with the gluster so I need to check this with our gluster
 gurus.​

​==​



 Thanks,
 Vijay

___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] release-3.6 branch created

2014-07-18 Thread Anders Blomdell
On 2014-07-17 20:22, Vijay Bellur wrote:
 Hi All,
 
 A new branch, 'release-3.6', has been branched from this commit in
 master:
 
 commit 950f9d8abe714708ca62b86f304e7417127e1132 Author: Jeff Darcy
 jda...@redhat.com Date:   Tue Jul 8 21:56:04 2014 -0400
 
 dht: fix rename race
 
 
 You can checkout this branch through:
 
 $git checkout -b release-3.6 origin/release-3.6
 
 rfc.sh is being updated to send patches to the appropriate branch.
 The plan is to have all 3.6.x releases happen off this branch. If you
 need any fix to be part of a 3.4.x release, please send out a
 backport of the same from master to release-3.4 after it has been
 accepted in master. More notes on backporting are available at [1].

Shouldn't the root of this branch get a tag to avoid this weirdness:

  # git checkout -b release-3.6 origin/release-3.6
  Branch release-3.6 set up to track remote branch release-3.6 from origin.
  Switched to a new branch 'release-3.6'

  # git describe
  v3.5qa2-762-g950f9d8

or have I missed some git magic?

/Anders

-- 
Anders Blomdell  Email: anders.blomd...@control.lth.se
Department of Automatic Control
Lund University  Phone:+46 46 222 4625
P.O. Box 118 Fax:  +46 46 138118
SE-221 00 Lund, Sweden

___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] Duplicate entries and other weirdness in a 3*4 volume

2014-07-18 Thread Lalatendu Mohanty

On 07/18/2014 07:57 PM, Anders Blomdell wrote:

During testing of a 3*4 gluster (from master as of yesterday), I encountered
two major weirdnesses:

   1. A 'rm -rf some_dir' needed several invocations to finish, each time
  reporting a number of lines like these:

	rm: cannot remove ‘a/b/c/d/e/f’: Directory not empty


   2. After having successfully deleted all files from the volume,
  i have a single directory that is duplicated in gluster-fuse,
  like this:
# ls -l /mnt/gluster
 total 24
 drwxr-xr-x 2 root root 12288 18 jul 16.17 work2/
 drwxr-xr-x 2 root root 12288 18 jul 16.17 work2/

any idea on how to debug this issue?

/Anders
  


Anders,

Check Gluster log files present in /var/log/glusterfs. Specifically 
glusterd logfile i.e. 
/var/log/glusterfs/etc-glusterfs-glusterd.vol.log. You can also start 
glusterd with debug mode i.e. $glusterd -L DEBUG and check the log 
files for more information.


Thanks,
Lala
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] Duplicate entries and other weirdness in a 3*4 volume

2014-07-18 Thread Pranith Kumar Karampuri


On 07/18/2014 07:57 PM, Anders Blomdell wrote:

During testing of a 3*4 gluster (from master as of yesterday), I encountered
two major weirdnesses:

   1. A 'rm -rf some_dir' needed several invocations to finish, each time
  reporting a number of lines like these:

	rm: cannot remove ‘a/b/c/d/e/f’: Directory not empty


   2. After having successfully deleted all files from the volume,
  i have a single directory that is duplicated in gluster-fuse,
  like this:
# ls -l /mnt/gluster
 total 24
 drwxr-xr-x 2 root root 12288 18 jul 16.17 work2/
 drwxr-xr-x 2 root root 12288 18 jul 16.17 work2/

any idea on how to debug this issue?
What are the steps to recreate? We need to first find what lead to this. 
Then probably which xlator leads to this.


Pranith


/Anders
  


___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel