Re: [Gluster-devel] Need a way to display and flush gluster cache ?

2016-07-25 Thread Prashanth Pai
+1 to option (2) which similar to echoing into /proc/sys/vm/drop_caches -Prashanth Pai - Original Message - > From: "Mohammed Rafi K C" > To: "gluster-users" , "Gluster Devel" > > Sent: Tuesday, 26 July, 2016 10:44:15 AM > Subject: [Gluster-devel] Need a way to display and flush glust

Re: [Gluster-devel] Open Source Storage Summit

2016-07-25 Thread Kaushal M
On Mon, Jul 25, 2016 at 11:20 AM, Raghavendra Gowdappa wrote: > > > - Original Message - >> From: "Amye Scavarda" >> To: "Gluster Devel" , "OSAS" >> Sent: Saturday, July 23, 2016 3:13:20 AM >> Subject: [Gluster-devel] Open Source Storage Summit >> >> This was a little late being announce

Re: [Gluster-devel] regression failed : snapshot/bug-1316437.t

2016-07-25 Thread Atin Mukherjee
On Tue, Jul 26, 2016 at 12:34 AM, Niels de Vos wrote: > On Mon, Jul 25, 2016 at 04:34:17PM +0530, Avra Sengupta wrote: > > The crux of the problem is that as of today, brick processes on restart > try > > to reuse the old port they were using (assuming that no other process > will > > be using it

[Gluster-devel] Need a way to display and flush gluster cache ?

2016-07-25 Thread Mohammed Rafi K C
Hi, Gluster stack has it's own caching mechanism , mostly on client side. But there is no concrete method to see how much memory are consuming by gluster for caching and if needed there is no way to flush the cache memory. So my first question is, Do we require to implement this two features for

[Gluster-devel] Status of GlusterFS-3.6

2016-07-25 Thread Kaushal M
Hi all! The GlusterFS-3.6 series has been around for nearly 2 years now. The first 3.6.0 release happened late October 2014. We've had 9 releases since then, with latest release 3.6.9 happening in early March 2016. There hasn't been any activity on the release-3.6 branch since then. The 3.6 series

Re: [Gluster-devel] GF_PARENT_DOWN on SIGKILL

2016-07-25 Thread Jeff Darcy
> As far as I know, there's no explicit guarantee on the order in which > fini is called, so we cannot rely on it to do cleanup because ec needs > that all its underlying xlators be fully functional to finish the cleanup. What kind of cleanup are we talking about here? We already need to handle t

Re: [Gluster-devel] GF_PARENT_DOWN on SIGKILL

2016-07-25 Thread Vijay Bellur
On 07/25/2016 02:41 AM, Xavier Hernandez wrote: Hi Jeff, On 22/07/16 15:37, Jeff Darcy wrote: Gah! sorry sorry, I meant to send the mail as SIGTERM. Not SIGKILL. So xavi and I were wondering why cleanup_and_exit() is not sending GF_PARENT_DOWN event. OK, then that grinding sound you hear is m

Re: [Gluster-devel] regression failed : snapshot/bug-1316437.t

2016-07-25 Thread Niels de Vos
On Mon, Jul 25, 2016 at 04:34:17PM +0530, Avra Sengupta wrote: > The crux of the problem is that as of today, brick processes on restart try > to reuse the old port they were using (assuming that no other process will > be using it, and not consulting pmap_registry_alloc() before using it). With >

Re: [Gluster-devel] Support to reclaim locks (posix) provided lkowner & range matches

2016-07-25 Thread Vijay Bellur
On 07/22/2016 08:44 AM, Soumya Koduri wrote: Hi, In certain scenarios (esp.,in highly available environments), the application may have to fail-over/connect to a different glusterFS client while the I/O is happening. In such cases until there is a ping timer expiry and glusterFS server cleans up

Re: [Gluster-devel] Proposing a framework to leverage existing Python unit test standards for our testing

2016-07-25 Thread Jonathan Holloway
- Original Message - > From: "Amye Scavarda" > To: "Vijay Bellur" > Cc: "Kaushal M" , "Jonathan Holloway" > , "Gluster Devel" > Sent: Sunday, July 24, 2016 7:53:06 PM > Subject: Re: [Gluster-devel] Proposing a framework to leverage existing > Python unit test standards for our testing

Re: [Gluster-devel] readdir() harmful in threaded code

2016-07-25 Thread Niels de Vos
On Mon, Jul 25, 2016 at 05:43:30AM +0200, Emmanuel Dreyfus wrote: > Vijay Bellur wrote: > > > Do you have any concrete examples of problems encountered due to the > > same directory stream being invoked from multiple threads? > > I am not sure this scenario can happen, but what we had were dire

Re: [Gluster-devel] Reducing regression runs (hopefully)

2016-07-25 Thread Nigel Babu
On Mon, Jul 25, 2016 at 07:37:55PM +0530, Raghavendra Talur wrote: > > > > > > > > 1. We do not clear the Verified tag. This means if you want to re-run > >regressions you have to manually trigger it. If your patch is rebased > > on top > >of another patch, you may have to retrigger failing

Re: [Gluster-devel] Reducing regression runs (hopefully)

2016-07-25 Thread Raghavendra Talur
> > > > > On the more ambitious side, I think we could also optimize around the > idea > > that some parts of the system aren't even present for some tests. For > > example, most of our tests don't even use GFAPI and won't be affected by > a > > GFAPI-only change. Conversely, GFAPI tests won't be

Re: [Gluster-devel] Reducing regression runs (hopefully)

2016-07-25 Thread Raghavendra Talur
On Mon, Jul 25, 2016 at 6:08 PM, Jeff Darcy wrote: > > I have a few proposals to reduce this turn around time: > > > > 1. We do not clear the Verified tag. This means if you want to re-run > >regressions you have to manually trigger it. If your patch is rebased > on > >top > >of anoth

Re: [Gluster-devel] Reducing regression runs (hopefully)

2016-07-25 Thread Raghavendra Talur
> > > > 1. We do not clear the Verified tag. This means if you want to re-run >regressions you have to manually trigger it. If your patch is rebased > on top >of another patch, you may have to retrigger failing regressions > manually. > Yes, matching the settings for Verified flag to that o

Re: [Gluster-devel] Reducing regression runs (hopefully)

2016-07-25 Thread Jeff Darcy
> My vision in this regard is something like this: > * A patchset gets Verified +1. > * A meta job is kicked off which determines regression jobs to run. > If the patch only touches GFAPI, we kick off the GFAPI regression tests. If > it touches multiple modules, we kick off the tests for these

Re: [Gluster-devel] regression failed : snapshot/bug-1316437.t

2016-07-25 Thread Atin Mukherjee
On Mon, Jul 25, 2016 at 7:12 PM, Atin Mukherjee wrote: > > > On Mon, Jul 25, 2016 at 5:37 PM, Atin Mukherjee > wrote: > >> >> >> On Mon, Jul 25, 2016 at 4:34 PM, Avra Sengupta >> wrote: >> >>> The crux of the problem is that as of today, brick processes on restart >>> try to reuse the old port

Re: [Gluster-devel] regression failed : snapshot/bug-1316437.t

2016-07-25 Thread Atin Mukherjee
On Mon, Jul 25, 2016 at 5:37 PM, Atin Mukherjee wrote: > > > On Mon, Jul 25, 2016 at 4:34 PM, Avra Sengupta > wrote: > >> The crux of the problem is that as of today, brick processes on restart >> try to reuse the old port they were using (assuming that no other process >> will be using it, and

[Gluster-devel] NetBSD now with more space

2016-07-25 Thread Nigel Babu
Hello folks, With Emmanuel's help, I've added an extra partition to the NetBSD machines. The /build and /archives folders are now in a new partition mounted at /data. This partition has about 23 GB of space. It will reduce instances of NetBSD machines out of action because of lack of space. A hug

Re: [Gluster-devel] Reducing regression runs (hopefully)

2016-07-25 Thread Nigel Babu
On Mon, Jul 25, 2016 at 08:38:34AM -0400, Jeff Darcy wrote: > > I have a few proposals to reduce this turn around time: > > > > 1. We do not clear the Verified tag. This means if you want to re-run > >regressions you have to manually trigger it. If your patch is rebased on > >top > >of

Re: [Gluster-devel] Reducing regression runs (hopefully)

2016-07-25 Thread Jeff Darcy
> I have a few proposals to reduce this turn around time: > > 1. We do not clear the Verified tag. This means if you want to re-run >regressions you have to manually trigger it. If your patch is rebased on >top >of another patch, you may have to retrigger failing regressions manually.

Re: [Gluster-devel] regression failed : snapshot/bug-1316437.t

2016-07-25 Thread Atin Mukherjee
On Mon, Jul 25, 2016 at 4:34 PM, Avra Sengupta wrote: > The crux of the problem is that as of today, brick processes on restart > try to reuse the old port they were using (assuming that no other process > will be using it, and not consulting pmap_registry_alloc() before using > it). With a recen

Re: [Gluster-devel] readdir() harmful in threaded code

2016-07-25 Thread Kaleb S. KEITHLEY
On 07/25/2016 07:26 AM, Kaleb S. KEITHLEY wrote: > On 07/23/2016 10:32 AM, Emmanuel Dreyfus wrote: >> Pranith Kumar Karampuri wrote: >> >>> So should we do readdir() with external locks for everything instead? >> >> readdir() with a per-directory lock is safe. However, it may come with a >> perfor

Re: [Gluster-devel] readdir() harmful in threaded code

2016-07-25 Thread Kaleb S. KEITHLEY
On 07/23/2016 10:32 AM, Emmanuel Dreyfus wrote: > Pranith Kumar Karampuri wrote: > >> So should we do readdir() with external locks for everything instead? > > readdir() with a per-directory lock is safe. However, it may come with a > performance hit in some scenarios, since two threads cannot r

Re: [Gluster-devel] regression failed : snapshot/bug-1316437.t

2016-07-25 Thread Avra Sengupta
The crux of the problem is that as of today, brick processes on restart try to reuse the old port they were using (assuming that no other process will be using it, and not consulting pmap_registry_alloc() before using it). With a recent change, pmap_registry_alloc (), reassigns older ports that

[Gluster-devel] Reducing regression runs (hopefully)

2016-07-25 Thread Nigel Babu
Hello, I'm looking to reduce the number of times we run regressions. It takes a lot of time and often times, it's not needed. We carry over the regression flags on rebase[1] and no code change[2] in Gerrit settings. However, Verified flag is cleared. When we re-grant the verify, we run regression

Re: [Gluster-devel] Why rdma.so is missing often

2016-07-25 Thread Prashanth Pai
- Original Message - > From: "jayakrishnan mm" > To: "Gluster Devel" , gluster-us...@gluster.org > Sent: Monday, 25 July, 2016 3:27:42 PM > Subject: [Gluster-devel] Why rdma.so is missing often > > Gluster version 3.7.6 > I get "Connection failed. Please check if gluster daemon is operat

[Gluster-devel] Why rdma.so is missing often

2016-07-25 Thread jayakrishnan mm
Gluster version 3.7.6 I get "Connection failed. Please check if gluster daemon is operational.", sometimes. When I check the log file , it says that rdma.so cannot be found. I did remove all the vol files and restarted the glusterd. Again when I tried to create volume, the connectio

Re: [Gluster-devel] regression failed : snapshot/bug-1316437.t

2016-07-25 Thread Avra Sengupta
The failure suggests that the port snapd is trying to bind to is already in use. But snapd has been modified to use a new port everytime. I am looking into this. On 07/25/2016 02:23 PM, Nithya Balachandran wrote: More failures: https://build.gluster.org/job/rackspace-regression-2GB-triggered/2

Re: [Gluster-devel] What tranpsort type in glfs_set_volfile_server() exactly mean ?

2016-07-25 Thread Raghavendra G
On Mon, Jul 18, 2016 at 4:35 PM, Raghavendra Talur wrote: > > > On Mon, Jul 18, 2016 at 4:10 PM, Prasanna Kalever > wrote: > >> Hey Team, >> >> >> My understanding is that @transport argumet in >> glfs_set_volfile_server() is meant for specifying transport used in >> fetching volfile server, >>

Re: [Gluster-devel] regression failed : snapshot/bug-1316437.t

2016-07-25 Thread Nithya Balachandran
More failures: https://build.gluster.org/job/rackspace-regression-2GB-triggered/22452/console I see these messages in the snapd.log: [2016-07-22 05:31:52.482282] I [rpcsvc.c:2199:rpcsvc_set_outstanding_rpc_limit] 0-rpc-service: Configured rpc.outstanding-rpc-limit with value 64 [2016-07-22 05:31: