[Gluster-devel] gluster-libgfapi-qemu-libvirt work status

2016-02-09 Thread Prasanna Kumar Kalever
Hi Team,

Be notified that work on qemu - libvirt - libgfapi integration for the better 
availability of servers, is on its way

what I try to do ?

qemu: modified glustefs backend code to fit space for multiple volfile serves 
a.k.a the backup volile servers
https://lists.nongnu.org/archive/html/qemu-devel/2015-11/msg02959.html

libvirt: enhancement to the network disk support for gluster
https://www.redhat.com/archives/libvir-list/2015-October/msg00374.html


This patches are WIP, stuck from qemu side with QAPI design issues, hopefully 
will be out in a couple of months.

Stay tuned :)

Thanks,
-Prasanna ​ 


___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel

Re: [Gluster-devel] changelog bug

2016-02-09 Thread Kotresh Hiremath Ravishankar
Hi

I think the two crashes he is seeing are in changelog_rollver and 
changelog_notifier are related
but not the one in BZ 1221629

addr2line is mapping to below lines.
changelog_rollover + 0xa9FD_ZERO(&rset);
changelog_notifier + 0x3ee   if (FD_ISSET (cn->rfd, &rset)) {

I am guessing it could be related to FD_SETSIZE and number of open file 
descriptors?


Thanks and Regards,
Kotresh H R

- Original Message -
> From: "Joe Julian" 
> To: gluster-devel@gluster.org
> Sent: Wednesday, February 10, 2016 10:18:07 AM
> Subject: Re: [Gluster-devel] changelog bug
> 
> btw... he was also having another crash in changelog_rollover:
> https://gist.githubusercontent.com/CyrilPeponnet/11954cbca725d4b8da7a/raw/2168169f7b208d8ee6193c4a444639505efb634b/gistfile1.txt
> 
> It would be a pretty huge coincidence if these were all unique causes,
> wouldn't it?
> 
> On 02/09/2016 08:27 PM, Kotresh Hiremath Ravishankar wrote:
> > Hi,
> >
> > This crash can't be same as BZ 1221629. The crash in the BZ 1221629
> > is with the rpc introduced in changelog in 3.7 along with bitrot.
> > Could you share the crash dump to analyse ?
> >
> > Thanks and Regards,
> > Kotresh H R
> >
> > - Original Message -
> >> From: "Vijay Bellur" 
> >> To: "Kotresh Hiremath Ravishankar" , "Manikandan
> >> Selvaganesh" 
> >> Cc: gluster-devel@gluster.org, "cyril peponnet"
> >> 
> >> Sent: Tuesday, February 9, 2016 10:26:57 PM
> >> Subject: Re: [Gluster-devel] changelog bug
> >>
> >> On 02/08/2016 01:14 AM, Kotresh Hiremath Ravishankar wrote:
> >>> Hi,
> >>>
> >>> This bug is already tracked BZ 1221629
> >>> I will start working on this and will update once it is fixed.
> >>>
> >> Cyril (in CC) also reported a similar crash with changelog in 3.6.5:
> >>
> >> https://gist.github.com/CyrilPeponnet/b67b360f186f31d34d8f
> >>
> >> The crash seems to be consistently reproducible in Cyril's setup. Can we
> >> address this soon?
> >>
> >> Thanks,
> >> Vijay
> >>
> >>
> > ___
> > Gluster-devel mailing list
> > Gluster-devel@gluster.org
> > http://www.gluster.org/mailman/listinfo/gluster-devel
> 
> ___
> Gluster-devel mailing list
> Gluster-devel@gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-devel
> 
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] Rebalance data migration and corruption

2016-02-09 Thread Soumya Koduri



On 02/09/2016 12:30 PM, Raghavendra G wrote:

   Right. But if there are simultaneous access to the same file from


 any other client and rebalance process, delegations shall
not be
 granted or revoked if granted even though they are operating at
 different offsets. So if you rely only on delegations,
migration may
 not proceed if an application has held a lock or doing any
I/Os.


Does the brick process wait for the response of delegation holder
(rebalance process here) before it wipes out the
delegation/locks? If
that's the case, rebalance process can complete one transaction of
(read, src) and (write, dst) before responding to a delegation
recall.
That way there is no starvation for both applications and rebalance
process (though this makes both of them slower, but that cannot
helped I
think).


yes. Brick process should wait for certain period before revoking
the delegations forcefully in case if it is not returned by the
client. Also if required (like done by NFS servers) we can choose to
increase this timeout value at run time if the client is diligently
flushing the data.


hmm.. I would prefer an infinite timeout. The only scenario where brick
process can forcefully flush leases would be connection lose with
rebalance process. The more scenarios where brick can flush leases
without knowledge of rebalance process, we open up more race-windows for
this bug to occur.

In fact at least in theory to be correct, rebalance process should
replay all the transactions that happened during the lease which got
flushed out by brick (after re-acquiring that lease). So, we would like
to avoid any such scenarios.

Btw, what is the necessity of timeouts? Is it an insurance against rogue
clients who won't respond back to lease recalls?
yes. It is to protect from rogue clients and prevent starvation of other 
clients.


In the current design, every lease is associated with lease-id (like 
lockowner in case of locks) and all the further fops (I/Os) have to be 
done using this lease-id. So in case if any fop comes to brick process 
with the lease-id of the lease which got flushed by the brick process, 
we can send special error and rebalance process can then replay all 
those fops. Will that be sufficient?


CCin Poornima who has been implementing it.


Thanks,
Soumya
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


[Gluster-devel] glusterfsd core on NetBSD (https://build.gluster.org/job/rackspace-netbsd7-regression-triggered/14139/consoleFull)

2016-02-09 Thread Soumya Koduri

Hi Emmanuel,

I see a core generated in this regression run though all the tests seem 
to have passed. I do not have a netbsd machine to analyze the core.

Could you please take a look and let me know what the issue could have been?

Thanks,
Soumya
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] presentation slides used at devconf

2016-02-09 Thread Atin Mukherjee
It'd be better if you can send a PR to glusterdocs with the odp.

On 02/10/2016 11:55 AM, Prasanna Kumar Kalever wrote:
> Hi Team,
> 
> Here is the link to talk at devconf http://redhat.slides.com/pkalever/deck
> 
> Note: 1. use up and down arrows along with the left and right arrows for 
> navigation
>   2. click on video logos in the slides to watch them.
> 
> -Prasanna ​ 
> 
> 
> ___
> Gluster-devel mailing list
> Gluster-devel@gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-devel
> 
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel

[Gluster-devel] presentation slides used at devconf

2016-02-09 Thread Prasanna Kumar Kalever
Hi Team,

Here is the link to talk at devconf http://redhat.slides.com/pkalever/deck

Note: 1. use up and down arrows along with the left and right arrows for 
navigation
  2. click on video logos in the slides to watch them.

-Prasanna ​ 


___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel

Re: [Gluster-devel] crash in dht in https://build.gluster.org/job/rackspace-regression-2GB-triggered/18134/consoleFull

2016-02-09 Thread Raghavendra Gowdappa


- Original Message -
> From: "Raghavendra Gowdappa" 
> To: "Pranith Kumar Karampuri" 
> Cc: "Gluster Devel" , "Nithya Balachandran" 
> 
> Sent: Wednesday, February 10, 2016 10:16:45 AM
> Subject: Re: crash in dht in 
> https://build.gluster.org/job/rackspace-regression-2GB-triggered/18134/consoleFull
> 
> Seems like a double unwind (or frame corruption in tier)
> (gdb) down
> #5  0x7f3f8c34c0ac in dht_lookup_dir_cbk (frame=0x7f3f6c002a0c,
> cookie=0x7f3f6c0064bc, this=0x7f3f8803bad0, op_ret=-1, op_errno=2,
> inode=0x0,
> stbuf=0x7f3f6c0069d4, xattr=0x0, postparent=0x7f3f6c006c04)
> at
> 
> /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-common.c:724
> 724   ret = dht_selfheal_directory (frame, dht_lookup_selfheal_cbk,
> (gdb) p *frame
> $13 = {root = 0x29f000, parent = 0xadc0de7f3f6c, frames = {next =
> 0x50e40de, prev = 0x307f3f88}, local = 0xac00,
>   this = 0x29f000, ret = 0xadc0de7f3f6c, ref_count = 222, lock =
>   84819968, cookie = 0x307f3f88, complete = _gf_false, op = 44032, begin
>   = {
> tv_sec = 11804356835803136, tv_usec = -5926493018029867156}, end =
> {tv_sec = 364298988607766750, tv_usec = 52776566472584},
>   wind_from = 0xac00 , wind_to
>   = 0x29f000 ,
>   unwind_from = 0xff7f3f6cbounds>, unwind_to = 0xbounds>}
> (gdb) p this->name
> $14 = 0x7f3f8803b640 "patchy-tier-dht"

(gdb) up
#6  0x7f3f8fb5ea98 in dht_lookup_dir_cbk (frame=0x7f3f6c0064bc, 
cookie=0x7f3f6c0084dc, this=0x7f3f8803aa80, op_ret=-1, op_errno=2, 
inode=0x7f3f6c0013ec, 
stbuf=0x7f3f94bc2860, xattr=0x0, postparent=0x7f3f94bc27f0)
at 
/home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-common.c:714
714 DHT_STACK_UNWIND (lookup, frame, local->op_ret, 
local->op_errno,
(gdb) p this->name
$22 = 0x7f3f88039510 "patchy-hot-dht"
(gdb) p frame->parent
$23 = (call_frame_t *) 0xadc0de7f3f6c
(gdb) p *frame->parent
Cannot access memory at address 0xadc0de7f3f6c
(gdb) p *frame
$24 = {root = 0x64a000, parent = 0xadc0de7f3f6c, frames = {next = 
0x50e40de, prev = 0x307f3f88}, local = 0xac00, 
  this = 0x64a000, ret = 0xadc0de7f3f6c, ref_count = 222, lock = 
84819968, cookie = 0x307f3f88, complete = _gf_false, op = 44032, begin = {
tv_sec = 28323419531509760, tv_usec = -5926493018029867156}, end = {tv_sec 
= 364298988607766750, tv_usec = 52776566472584}, 
  wind_from = 0xac00 , wind_to = 
0x64a000 , 
  unwind_from = 0xff7f3f6c , 
unwind_to = 0x }
(gdb) p this->name
$25 = 0x7f3f88039510 "patchy-hot-dht"
(gdb) up
#7  0x7f3f8fddb23a in client3_3_lookup_cbk (req=0x7f3f6c0090ac, 
iov=0x7f3f6c0090ec, count=1, myframe=0x7f3f6c0084dc)
at 
/home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/protocol/client/src/client-rpc-fops.c:3028
3028
/home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/protocol/client/src/client-rpc-fops.c:
 No such file or directory.
(gdb) p frame
$26 = (call_frame_t *) 0x7f3f6c0084dc
(gdb) p *frame
$27 = {root = 0x84c000, parent = 0xadc0de7f3f6c, frames = {next = 
0x54d20de, prev = 0x307f3f88}, local = 0xac00, 
  this = 0x84c000, ret = 0xadc0de7f3f6c, ref_count = 222, lock = 
88940544, cookie = 0x307f3f88, complete = _gf_false, op = 44032, begin = {
tv_sec = 37365803158339584, tv_usec = -5926493018029867156}, end = {tv_sec 
= 381996727768449246, tv_usec = 52776566472584}, 
  wind_from = 0xac00 , wind_to = 
0x84c000 , 
  unwind_from = 0xff7f3f6c , 
unwind_to = 0x }
(gdb) p this->name
$28 = 0x7f3f880202d0 "patchy-client-5"

Seems like entire stack is corrupted

> No, I am not aware of this crash.
> 
> - Original Message -
> > From: "Pranith Kumar Karampuri" 
> > To: "Gluster Devel" , "Raghavendra Gowdappa"
> > , "Nithya Balachandran"
> > 
> > Sent: Wednesday, February 10, 2016 5:51:32 AM
> > Subject: crash in dht in
> > https://build.gluster.org/job/rackspace-regression-2GB-triggered/18134/consoleFull
> > 
> > hi,
> > I see the following crash. Is this a known issue?
> > (gdb) bt
> > #0  0x7f3f8c339fb4 in dht_selfheal_dir_setattr
> > (frame=0x7f3f6c002a0c, loc=0x7f3f6c000944, stbuf=0x7f3f6c0009d4,
> > valid=16777215,
> >  layout=0x7f3f6c004140) at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-selfheal.c:1087
> > #1  0x7f3f8c33a4f8 in dht_selfheal_dir_mkdir_cbk
> > (frame=0x7f3f6c002a0c, cookie=0x7f3f9004201c, this=0x7f3f8803bad0,
> > op_ret=-1, op_errno=5,
> >  inode=0x0, stbuf=0x0, preparent=0x0, postparent=0x0, xdata=0x0)
> >  at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-selfheal.c:1147
> > #2  0x7

Re: [Gluster-devel] changelog bug

2016-02-09 Thread Joe Julian
btw... he was also having another crash in changelog_rollover: 
https://gist.githubusercontent.com/CyrilPeponnet/11954cbca725d4b8da7a/raw/2168169f7b208d8ee6193c4a444639505efb634b/gistfile1.txt


It would be a pretty huge coincidence if these were all unique causes, 
wouldn't it?


On 02/09/2016 08:27 PM, Kotresh Hiremath Ravishankar wrote:

Hi,

This crash can't be same as BZ 1221629. The crash in the BZ 1221629
is with the rpc introduced in changelog in 3.7 along with bitrot.
Could you share the crash dump to analyse ?

Thanks and Regards,
Kotresh H R

- Original Message -

From: "Vijay Bellur" 
To: "Kotresh Hiremath Ravishankar" , "Manikandan Selvaganesh" 

Cc: gluster-devel@gluster.org, "cyril peponnet" 

Sent: Tuesday, February 9, 2016 10:26:57 PM
Subject: Re: [Gluster-devel] changelog bug

On 02/08/2016 01:14 AM, Kotresh Hiremath Ravishankar wrote:

Hi,

This bug is already tracked BZ 1221629
I will start working on this and will update once it is fixed.


Cyril (in CC) also reported a similar crash with changelog in 3.6.5:

https://gist.github.com/CyrilPeponnet/b67b360f186f31d34d8f

The crash seems to be consistently reproducible in Cyril's setup. Can we
address this soon?

Thanks,
Vijay



___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] crash in dht in https://build.gluster.org/job/rackspace-regression-2GB-triggered/18134/consoleFull

2016-02-09 Thread Raghavendra Gowdappa
Seems like a double unwind (or frame corruption in tier)
(gdb) down
#5  0x7f3f8c34c0ac in dht_lookup_dir_cbk (frame=0x7f3f6c002a0c, 
cookie=0x7f3f6c0064bc, this=0x7f3f8803bad0, op_ret=-1, op_errno=2, inode=0x0, 
stbuf=0x7f3f6c0069d4, xattr=0x0, postparent=0x7f3f6c006c04)
at 
/home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-common.c:724
724 ret = dht_selfheal_directory (frame, dht_lookup_selfheal_cbk,
(gdb) p *frame
$13 = {root = 0x29f000, parent = 0xadc0de7f3f6c, frames = {next = 
0x50e40de, prev = 0x307f3f88}, local = 0xac00, 
  this = 0x29f000, ret = 0xadc0de7f3f6c, ref_count = 222, lock = 
84819968, cookie = 0x307f3f88, complete = _gf_false, op = 44032, begin = {
tv_sec = 11804356835803136, tv_usec = -5926493018029867156}, end = {tv_sec 
= 364298988607766750, tv_usec = 52776566472584}, 
  wind_from = 0xac00 , wind_to = 
0x29f000 , 
  unwind_from = 0xff7f3f6c , 
unwind_to = 0x }
(gdb) p this->name
$14 = 0x7f3f8803b640 "patchy-tier-dht"

No, I am not aware of this crash.

- Original Message -
> From: "Pranith Kumar Karampuri" 
> To: "Gluster Devel" , "Raghavendra Gowdappa" 
> , "Nithya Balachandran"
> 
> Sent: Wednesday, February 10, 2016 5:51:32 AM
> Subject: crash in dht in 
> https://build.gluster.org/job/rackspace-regression-2GB-triggered/18134/consoleFull
> 
> hi,
> I see the following crash. Is this a known issue?
> (gdb) bt
> #0  0x7f3f8c339fb4 in dht_selfheal_dir_setattr
> (frame=0x7f3f6c002a0c, loc=0x7f3f6c000944, stbuf=0x7f3f6c0009d4,
> valid=16777215,
>  layout=0x7f3f6c004140) at
> /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-selfheal.c:1087
> #1  0x7f3f8c33a4f8 in dht_selfheal_dir_mkdir_cbk
> (frame=0x7f3f6c002a0c, cookie=0x7f3f9004201c, this=0x7f3f8803bad0,
> op_ret=-1, op_errno=5,
>  inode=0x0, stbuf=0x0, preparent=0x0, postparent=0x0, xdata=0x0)
>  at
> /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-selfheal.c:1147
> #2  0x7f3f8fb81f31 in dht_mkdir (frame=0x7f3f9004201c,
> this=0x7f3f8803aa80, loc=0x7f3f6c000944, mode=16877, umask=0,
> params=0x7f3f9003594c)
>  at
> /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-common.c:6710
> #3  0x7f3f8c33ad1f in dht_selfheal_dir_mkdir (frame=0x7f3f6c002a0c,
> loc=0x7f3f6c000944, layout=0x7f3f6c004140, force=0)
>  at
> /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-selfheal.c:1256
> #4  0x7f3f8c33c4c8 in dht_selfheal_directory (frame=0x7f3f6c002a0c,
> dir_cbk=0x7f3f8c349588 , loc=0x7f3f6c000944,
>  layout=0x7f3f6c004140) at
> /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-selfheal.c:1837
> #5  0x7f3f8c34c0ac in dht_lookup_dir_cbk (frame=0x7f3f6c002a0c,
> cookie=0x7f3f6c0064bc, this=0x7f3f8803bad0, op_ret=-1, op_errno=2,
> inode=0x0,
>  stbuf=0x7f3f6c0069d4, xattr=0x0, postparent=0x7f3f6c006c04)
>  at
> /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-common.c:724
> #6  0x7f3f8fb5ea98 in dht_lookup_dir_cbk (frame=0x7f3f6c0064bc,
> cookie=0x7f3f6c0084dc, this=0x7f3f8803aa80, op_ret=-1, op_errno=2,
>  inode=0x7f3f6c0013ec, stbuf=0x7f3f94bc2860, xattr=0x0,
> postparent=0x7f3f94bc27f0)
>  at
> /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-common.c:714
> #7  0x7f3f8fddb23a in client3_3_lookup_cbk (req=0x7f3f6c0090ac,
> iov=0x7f3f6c0090ec, count=1, myframe=0x7f3f6c0084dc)
>  at
> /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/protocol/client/src/client-rpc-fops.c:3028
> #8  0x7f3fa0c2d42a in rpc_clnt_handle_reply (clnt=0x7f3f8806cdb0,
> pollin=0x7f3f90041c90)
>  at
> /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-clnt.c:759
> #9  0x7f3fa0c2d8c8 in rpc_clnt_notify (trans=0x7f3f8806d240,
> mydata=0x7f3f8806cde0, event=RPC_TRANSPORT_MSG_RECEIVED,
> data=0x7f3f90041c90)
>  at
> /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-clnt.c:900
> #10 0x7f3fa0c29b5a in rpc_transport_notify (this=0x7f3f8806d240,
> event=RPC_TRANSPORT_MSG_RECEIVED, data=0x7f3f90041c90)
>  at
> /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-transport.c:541
> #11 0x7f3f961eadcb in socket_event_poll_in (this=0x7f3f8806d240)
>  at
> /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-transport/socket/src/socket.c:2231
> #12 0x7f3f961eb321 in socket_event_handler (fd=18, idx=12,
> data=0x7f3f8806d240, poll_in=1, poll_out=0, poll_err=0)
>  at
> /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-transport/socket/src/socket.c:2344
> 

Re: [Gluster-devel] Gluster test infrastructure and current challenges

2016-02-09 Thread Raghavendra Talur
On Tue, Feb 9, 2016 at 9:12 PM, Amye Scavarda  wrote:

>
>
> On Tue, Feb 9, 2016 at 3:41 PM, Raghavendra Talur 
> wrote:
>
>> Hey folks,
>>
>> Gluster test infrastructure needs a bit of attention. We have committed a
>> lot of code in the last couple of years, but we have not scaled our
>> infrastructure at the same rate. Lately, the signs have become alarming and
>> it calls for attention.
>>
>> We have scheduled a hangout[1] to educate new Devs about
>> 1. Current Gluster test framework - TAP, prove, include.rc.
>> 2. Challenges
>>  a. Regression takes a lot of time to run
>>  b. We have so many non-deterministic tests
>>  c. Identifying cause  for test failure takes time
>>  d. NetBSD debugging knowledge in the community is scarce.
>>  e. DISTAF: Multinode testing is not integrated with regression yet.
>>  f. More types of tests to have: upgrade, performance, integration.
>>  g. unit tests
>>
>> Note that agenda is mainly to showcase the current challenges through a
>> medium more informative than email/irc. It is not to discuss possible
>> solutions; that should be done over email on the gluster-devel mailing list.
>>
>> Hence this hangout is scheduled keeping in view convenience of devs from
>> IST or around. We will have the session recorded for everyone to view at
>> their own convenience. Also we could do the same session for different time
>> zones.
>>
>
> This is for tomorrow, Feb 10th? Or am I misreading this event?
> With this short of notice, will we be able to get people to attend?
>
> Thanks!
>  -amye
>
>


Originally this was supposed to be a record and upload to youtube kind of
presentation.
We changed it to hangout thinking it could help being a realtime broadcast.

Considering it now, I feel we add no value to the event making it a hangout
if it is just a demo.
We need a hangout/irc meeting for discussion.

I will cancel the hangout for now and upload a demo somewhere with the
agenda mentioned above.(It would just explain current state)
We can discuss in community meeting today for a better date and time to
hold a meeting to discuss solutions for challenges that we mention in the
demo.

Thanks for the input Amye.

NOTE: The hangout has been cancelled. We will send invite for the
rescheduled session later.


>>
>> [1] https://plus.google.com/events/c25hcj9llrvvp3dcidqk26sjrhs
>>
>> Thanks,
>> Raghavendra Talur
>>
>> ___
>> Gluster-devel mailing list
>> Gluster-devel@gluster.org
>> http://www.gluster.org/mailman/listinfo/gluster-devel
>>
>
>
>
> --
> Amye Scavarda | a...@redhat.com | Gluster Community Lead
>
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel

Re: [Gluster-devel] changelog bug

2016-02-09 Thread Kotresh Hiremath Ravishankar
Hi,

This crash can't be same as BZ 1221629. The crash in the BZ 1221629
is with the rpc introduced in changelog in 3.7 along with bitrot.
Could you share the crash dump to analyse ?

Thanks and Regards,
Kotresh H R

- Original Message -
> From: "Vijay Bellur" 
> To: "Kotresh Hiremath Ravishankar" , "Manikandan 
> Selvaganesh" 
> Cc: gluster-devel@gluster.org, "cyril peponnet" 
> 
> Sent: Tuesday, February 9, 2016 10:26:57 PM
> Subject: Re: [Gluster-devel] changelog bug
> 
> On 02/08/2016 01:14 AM, Kotresh Hiremath Ravishankar wrote:
> > Hi,
> >
> > This bug is already tracked BZ 1221629
> > I will start working on this and will update once it is fixed.
> >
> 
> Cyril (in CC) also reported a similar crash with changelog in 3.6.5:
> 
> https://gist.github.com/CyrilPeponnet/b67b360f186f31d34d8f
> 
> The crash seems to be consistently reproducible in Cyril's setup. Can we
> address this soon?
> 
> Thanks,
> Vijay
> 
> 
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


[Gluster-devel] crash in dht in https://build.gluster.org/job/rackspace-regression-2GB-triggered/18134/consoleFull

2016-02-09 Thread Pranith Kumar Karampuri

hi,
   I see the following crash. Is this a known issue?
(gdb) bt
#0  0x7f3f8c339fb4 in dht_selfheal_dir_setattr 
(frame=0x7f3f6c002a0c, loc=0x7f3f6c000944, stbuf=0x7f3f6c0009d4, 
valid=16777215,
layout=0x7f3f6c004140) at 
/home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-selfheal.c:1087
#1  0x7f3f8c33a4f8 in dht_selfheal_dir_mkdir_cbk 
(frame=0x7f3f6c002a0c, cookie=0x7f3f9004201c, this=0x7f3f8803bad0, 
op_ret=-1, op_errno=5,

inode=0x0, stbuf=0x0, preparent=0x0, postparent=0x0, xdata=0x0)
at 
/home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-selfheal.c:1147
#2  0x7f3f8fb81f31 in dht_mkdir (frame=0x7f3f9004201c, 
this=0x7f3f8803aa80, loc=0x7f3f6c000944, mode=16877, umask=0, 
params=0x7f3f9003594c)
at 
/home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-common.c:6710
#3  0x7f3f8c33ad1f in dht_selfheal_dir_mkdir (frame=0x7f3f6c002a0c, 
loc=0x7f3f6c000944, layout=0x7f3f6c004140, force=0)
at 
/home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-selfheal.c:1256
#4  0x7f3f8c33c4c8 in dht_selfheal_directory (frame=0x7f3f6c002a0c, 
dir_cbk=0x7f3f8c349588 , loc=0x7f3f6c000944,
layout=0x7f3f6c004140) at 
/home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-selfheal.c:1837
#5  0x7f3f8c34c0ac in dht_lookup_dir_cbk (frame=0x7f3f6c002a0c, 
cookie=0x7f3f6c0064bc, this=0x7f3f8803bad0, op_ret=-1, op_errno=2, 
inode=0x0,

stbuf=0x7f3f6c0069d4, xattr=0x0, postparent=0x7f3f6c006c04)
at 
/home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-common.c:724
#6  0x7f3f8fb5ea98 in dht_lookup_dir_cbk (frame=0x7f3f6c0064bc, 
cookie=0x7f3f6c0084dc, this=0x7f3f8803aa80, op_ret=-1, op_errno=2,
inode=0x7f3f6c0013ec, stbuf=0x7f3f94bc2860, xattr=0x0, 
postparent=0x7f3f94bc27f0)
at 
/home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-common.c:714
#7  0x7f3f8fddb23a in client3_3_lookup_cbk (req=0x7f3f6c0090ac, 
iov=0x7f3f6c0090ec, count=1, myframe=0x7f3f6c0084dc)
at 
/home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/protocol/client/src/client-rpc-fops.c:3028
#8  0x7f3fa0c2d42a in rpc_clnt_handle_reply (clnt=0x7f3f8806cdb0, 
pollin=0x7f3f90041c90)
at 
/home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-clnt.c:759
#9  0x7f3fa0c2d8c8 in rpc_clnt_notify (trans=0x7f3f8806d240, 
mydata=0x7f3f8806cde0, event=RPC_TRANSPORT_MSG_RECEIVED, 
data=0x7f3f90041c90)
at 
/home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-clnt.c:900
#10 0x7f3fa0c29b5a in rpc_transport_notify (this=0x7f3f8806d240, 
event=RPC_TRANSPORT_MSG_RECEIVED, data=0x7f3f90041c90)
at 
/home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-transport.c:541

#11 0x7f3f961eadcb in socket_event_poll_in (this=0x7f3f8806d240)
at 
/home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-transport/socket/src/socket.c:2231
#12 0x7f3f961eb321 in socket_event_handler (fd=18, idx=12, 
data=0x7f3f8806d240, poll_in=1, poll_out=0, poll_err=0)
at 
/home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-transport/socket/src/socket.c:2344
#13 0x7f3fa0ec61a8 in event_dispatch_epoll_handler 
(event_pool=0x121fce0, event=0x7f3f94bc2e70)
at 
/home/jenkins/root/workspace/rackspace-regression-2GB-triggered/libglusterfs/src/event-epoll.c:571

#14 0x7f3fa0ec6596 in event_dispatch_epoll_worker (data=0x125e1a0)
at 
/home/jenkins/root/workspace/rackspace-regression-2GB-triggered/libglusterfs/src/event-epoll.c:674

#15 0x7f3fa0144a51 in start_thread () from ./lib64/libpthread.so.0
#16 0x7f3f9faae93d in clone () from ./lib64/libc.so.6


Pranith
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] changelog bug

2016-02-09 Thread Vijay Bellur

On 02/08/2016 01:14 AM, Kotresh Hiremath Ravishankar wrote:

Hi,

This bug is already tracked BZ 1221629
I will start working on this and will update once it is fixed.



Cyril (in CC) also reported a similar crash with changelog in 3.6.5:

https://gist.github.com/CyrilPeponnet/b67b360f186f31d34d8f

The crash seems to be consistently reproducible in Cyril's setup. Can we 
address this soon?


Thanks,
Vijay

___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] Gluster test infrastructure and current challenges

2016-02-09 Thread Amye Scavarda
On Tue, Feb 9, 2016 at 3:41 PM, Raghavendra Talur  wrote:

> Hey folks,
>
> Gluster test infrastructure needs a bit of attention. We have committed a
> lot of code in the last couple of years, but we have not scaled our
> infrastructure at the same rate. Lately, the signs have become alarming and
> it calls for attention.
>
> We have scheduled a hangout[1] to educate new Devs about
> 1. Current Gluster test framework - TAP, prove, include.rc.
> 2. Challenges
>  a. Regression takes a lot of time to run
>  b. We have so many non-deterministic tests
>  c. Identifying cause  for test failure takes time
>  d. NetBSD debugging knowledge in the community is scarce.
>  e. DISTAF: Multinode testing is not integrated with regression yet.
>  f. More types of tests to have: upgrade, performance, integration.
>  g. unit tests
>
> Note that agenda is mainly to showcase the current challenges through a
> medium more informative than email/irc. It is not to discuss possible
> solutions; that should be done over email on the gluster-devel mailing list.
>
> Hence this hangout is scheduled keeping in view convenience of devs from
> IST or around. We will have the session recorded for everyone to view at
> their own convenience. Also we could do the same session for different time
> zones.
>

This is for tomorrow, Feb 10th? Or am I misreading this event?
With this short of notice, will we be able to get people to attend?

Thanks!
 -amye


>
>
> [1] https://plus.google.com/events/c25hcj9llrvvp3dcidqk26sjrhs
>
> Thanks,
> Raghavendra Talur
>
> ___
> Gluster-devel mailing list
> Gluster-devel@gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-devel
>



-- 
Amye Scavarda | a...@redhat.com | Gluster Community Lead
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel

[Gluster-devel] Gluster test infrastructure and current challenges

2016-02-09 Thread Raghavendra Talur
Hey folks,

Gluster test infrastructure needs a bit of attention. We have committed a
lot of code in the last couple of years, but we have not scaled our
infrastructure at the same rate. Lately, the signs have become alarming and
it calls for attention.

We have scheduled a hangout[1] to educate new Devs about
1. Current Gluster test framework - TAP, prove, include.rc.
2. Challenges
 a. Regression takes a lot of time to run
 b. We have so many non-deterministic tests
 c. Identifying cause  for test failure takes time
 d. NetBSD debugging knowledge in the community is scarce.
 e. DISTAF: Multinode testing is not integrated with regression yet.
 f. More types of tests to have: upgrade, performance, integration.
 g. unit tests

Note that agenda is mainly to showcase the current challenges through a
medium more informative than email/irc. It is not to discuss possible
solutions; that should be done over email on the gluster-devel mailing list.

Hence this hangout is scheduled keeping in view convenience of devs from
IST or around. We will have the session recorded for everyone to view at
their own convenience. Also we could do the same session for different time
zones.

[1] https://plus.google.com/events/c25hcj9llrvvp3dcidqk26sjrhs

Thanks,
Raghavendra Talur
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel

[Gluster-devel] Minutes of today's Gluster Community Bug Triage Meeting

2016-02-09 Thread Manikandan Selvaganesh
Hi all,

Thanks everyone for attending the Gluster Community Bug Triage meeting today
and here are the minutes of the meeting:

Meeting summary
---
* agenda: https://public.pad.fsfe.org/p/gluster-bug-triage (Manikandan,
  12:00:41)
* Roll Call  (Manikandan, 12:00:48)

* kkeithley_ will come up with a proposal to reduce the number of bugs
  against "mainline" in NEW state  (Manikandan, 12:05:23)
  * ACTION: kkeithley_ will come up with a proposal to reduce the number
of bugs against "mainline" in NEW state  (Manikandan, 12:06:15)

* msvbhat  and ndevos need to think about and decide how to provide/use
  debug builds  (Manikandan, 12:06:35)
  * ACTION: msvbhat  and ndevos need to think about and decide how to
provide/use debug build  (Manikandan, 12:07:34)

* Group Triage  (Manikandan, 12:08:38)
  * LINK:
http://gluster.readthedocs.org/en/latest/Contributors-Guide/Bug-Triage/
(Manikandan, 12:09:07)

* Open Floor  (Manikandan, 12:26:23)

Meeting ended at 12:29:30 UTC.


Action Items

* kkeithley_ will come up with a proposal to reduce the number of bugs
  against "mainline" in NEW state
* msvbhat  and ndevos need to think about and decide how to provide/use
  debug build.
* kkeithley_ will come up with a proposal to reduce the number of bugs
  against "mainline" in NEW state
* hagarth start/sync email on regular (nightly) automated tests
* msvbhat will look into using nightly builds for automated testing,
  and will report issues/success to the mailinglist
* msvbhat will look into lalatenduM's automated Coverity setup in Jenkins
  whichneed assistance  from an admin with more permissions 
* msvbhat  and ndevos need to think about and decide how to provide/use
  debug builds
* msvbhat  provide a simple step/walk-through on how to provide testcases
  for the nightly rpm tests
* ndevos to propose some test-cases for minimal libgfapi test
* Manikandan and Nandaja will keep updating on the bug automation workflow.

People Present (lines said)
---
* Manikandan (43)
* ndevos (6)
* hgowtham (5)
* skoduri (4)
* jiffin (3)
* zodbot (3)
* Saravanakmr (3)

See you all next week :-)

--
Thanks & Regards,
Manikandan Selvaganesh.

___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


[Gluster-devel] REMINDER: Gluster Community Bug Triage meeting at 12:00 UTC (~in 60 minutes)

2016-02-09 Thread Manikandan Selvaganesh
Hi all,

This meeting is scheduled for anyone that is interested in learning more
about, or assisting with the Bug Triage.

Meeting details:
- location: #gluster-meeting on Freenode IRC
 (https://webchat.freenode.net/?channels=gluster-meeting  )
- date: every Tuesday
- time: 12:00 UTC
 (in your terminal, run: date -d "12:00 UTC")
- agenda: https://public.pad.fsfe.org/p/gluster-bug-triage

Currently the following items are listed:
* Roll Call
* Status of last weeks action items
* Group Triage
* Open Floor

The last two topics have space for additions. If you have a suitable bug
or topic to discuss, please add it to the agenda.

Appreciate your participation.

Thank you :-)

--
Regards,
Manikandan Selvaganesh.

___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] Cores on NetBSD of brick https://build.gluster.org/job/rackspace-netbsd7-regression-triggered/14100/consoleFull

2016-02-09 Thread Pranith Kumar Karampuri



On 02/09/2016 04:13 PM, Emmanuel Dreyfus wrote:

On Tue, Feb 09, 2016 at 11:56:37AM +0530, Pranith Kumar Karampuri wrote:

I think the regression run is not giving that link anymore when the crash
happens? Could you please add that also as a link in regression run?

Ther was the path of the archive, I changed it for a http:// link

Oops, sorry, it is there.

Pranith




___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] Cores on NetBSD of brick https://build.gluster.org/job/rackspace-netbsd7-regression-triggered/14100/consoleFull

2016-02-09 Thread Emmanuel Dreyfus
On Tue, Feb 09, 2016 at 11:56:37AM +0530, Pranith Kumar Karampuri wrote:
> I think the regression run is not giving that link anymore when the crash
> happens? Could you please add that also as a link in regression run?

Ther was the path of the archive, I changed it for a http:// link

-- 
Emmanuel Dreyfus
m...@netbsd.org
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel