[Gluster-devel] gluster-libgfapi-qemu-libvirt work status
Hi Team, Be notified that work on qemu - libvirt - libgfapi integration for the better availability of servers, is on its way what I try to do ? qemu: modified glustefs backend code to fit space for multiple volfile serves a.k.a the backup volile servers https://lists.nongnu.org/archive/html/qemu-devel/2015-11/msg02959.html libvirt: enhancement to the network disk support for gluster https://www.redhat.com/archives/libvir-list/2015-October/msg00374.html This patches are WIP, stuck from qemu side with QAPI design issues, hopefully will be out in a couple of months. Stay tuned :) Thanks, -Prasanna ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] changelog bug
Hi I think the two crashes he is seeing are in changelog_rollver and changelog_notifier are related but not the one in BZ 1221629 addr2line is mapping to below lines. changelog_rollover + 0xa9FD_ZERO(&rset); changelog_notifier + 0x3ee if (FD_ISSET (cn->rfd, &rset)) { I am guessing it could be related to FD_SETSIZE and number of open file descriptors? Thanks and Regards, Kotresh H R - Original Message - > From: "Joe Julian" > To: gluster-devel@gluster.org > Sent: Wednesday, February 10, 2016 10:18:07 AM > Subject: Re: [Gluster-devel] changelog bug > > btw... he was also having another crash in changelog_rollover: > https://gist.githubusercontent.com/CyrilPeponnet/11954cbca725d4b8da7a/raw/2168169f7b208d8ee6193c4a444639505efb634b/gistfile1.txt > > It would be a pretty huge coincidence if these were all unique causes, > wouldn't it? > > On 02/09/2016 08:27 PM, Kotresh Hiremath Ravishankar wrote: > > Hi, > > > > This crash can't be same as BZ 1221629. The crash in the BZ 1221629 > > is with the rpc introduced in changelog in 3.7 along with bitrot. > > Could you share the crash dump to analyse ? > > > > Thanks and Regards, > > Kotresh H R > > > > - Original Message - > >> From: "Vijay Bellur" > >> To: "Kotresh Hiremath Ravishankar" , "Manikandan > >> Selvaganesh" > >> Cc: gluster-devel@gluster.org, "cyril peponnet" > >> > >> Sent: Tuesday, February 9, 2016 10:26:57 PM > >> Subject: Re: [Gluster-devel] changelog bug > >> > >> On 02/08/2016 01:14 AM, Kotresh Hiremath Ravishankar wrote: > >>> Hi, > >>> > >>> This bug is already tracked BZ 1221629 > >>> I will start working on this and will update once it is fixed. > >>> > >> Cyril (in CC) also reported a similar crash with changelog in 3.6.5: > >> > >> https://gist.github.com/CyrilPeponnet/b67b360f186f31d34d8f > >> > >> The crash seems to be consistently reproducible in Cyril's setup. Can we > >> address this soon? > >> > >> Thanks, > >> Vijay > >> > >> > > ___ > > Gluster-devel mailing list > > Gluster-devel@gluster.org > > http://www.gluster.org/mailman/listinfo/gluster-devel > > ___ > Gluster-devel mailing list > Gluster-devel@gluster.org > http://www.gluster.org/mailman/listinfo/gluster-devel > ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] Rebalance data migration and corruption
On 02/09/2016 12:30 PM, Raghavendra G wrote: Right. But if there are simultaneous access to the same file from any other client and rebalance process, delegations shall not be granted or revoked if granted even though they are operating at different offsets. So if you rely only on delegations, migration may not proceed if an application has held a lock or doing any I/Os. Does the brick process wait for the response of delegation holder (rebalance process here) before it wipes out the delegation/locks? If that's the case, rebalance process can complete one transaction of (read, src) and (write, dst) before responding to a delegation recall. That way there is no starvation for both applications and rebalance process (though this makes both of them slower, but that cannot helped I think). yes. Brick process should wait for certain period before revoking the delegations forcefully in case if it is not returned by the client. Also if required (like done by NFS servers) we can choose to increase this timeout value at run time if the client is diligently flushing the data. hmm.. I would prefer an infinite timeout. The only scenario where brick process can forcefully flush leases would be connection lose with rebalance process. The more scenarios where brick can flush leases without knowledge of rebalance process, we open up more race-windows for this bug to occur. In fact at least in theory to be correct, rebalance process should replay all the transactions that happened during the lease which got flushed out by brick (after re-acquiring that lease). So, we would like to avoid any such scenarios. Btw, what is the necessity of timeouts? Is it an insurance against rogue clients who won't respond back to lease recalls? yes. It is to protect from rogue clients and prevent starvation of other clients. In the current design, every lease is associated with lease-id (like lockowner in case of locks) and all the further fops (I/Os) have to be done using this lease-id. So in case if any fop comes to brick process with the lease-id of the lease which got flushed by the brick process, we can send special error and rebalance process can then replay all those fops. Will that be sufficient? CCin Poornima who has been implementing it. Thanks, Soumya ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
[Gluster-devel] glusterfsd core on NetBSD (https://build.gluster.org/job/rackspace-netbsd7-regression-triggered/14139/consoleFull)
Hi Emmanuel, I see a core generated in this regression run though all the tests seem to have passed. I do not have a netbsd machine to analyze the core. Could you please take a look and let me know what the issue could have been? Thanks, Soumya ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] presentation slides used at devconf
It'd be better if you can send a PR to glusterdocs with the odp. On 02/10/2016 11:55 AM, Prasanna Kumar Kalever wrote: > Hi Team, > > Here is the link to talk at devconf http://redhat.slides.com/pkalever/deck > > Note: 1. use up and down arrows along with the left and right arrows for > navigation > 2. click on video logos in the slides to watch them. > > -Prasanna > > > ___ > Gluster-devel mailing list > Gluster-devel@gluster.org > http://www.gluster.org/mailman/listinfo/gluster-devel > ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
[Gluster-devel] presentation slides used at devconf
Hi Team, Here is the link to talk at devconf http://redhat.slides.com/pkalever/deck Note: 1. use up and down arrows along with the left and right arrows for navigation 2. click on video logos in the slides to watch them. -Prasanna ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] crash in dht in https://build.gluster.org/job/rackspace-regression-2GB-triggered/18134/consoleFull
- Original Message - > From: "Raghavendra Gowdappa" > To: "Pranith Kumar Karampuri" > Cc: "Gluster Devel" , "Nithya Balachandran" > > Sent: Wednesday, February 10, 2016 10:16:45 AM > Subject: Re: crash in dht in > https://build.gluster.org/job/rackspace-regression-2GB-triggered/18134/consoleFull > > Seems like a double unwind (or frame corruption in tier) > (gdb) down > #5 0x7f3f8c34c0ac in dht_lookup_dir_cbk (frame=0x7f3f6c002a0c, > cookie=0x7f3f6c0064bc, this=0x7f3f8803bad0, op_ret=-1, op_errno=2, > inode=0x0, > stbuf=0x7f3f6c0069d4, xattr=0x0, postparent=0x7f3f6c006c04) > at > > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-common.c:724 > 724 ret = dht_selfheal_directory (frame, dht_lookup_selfheal_cbk, > (gdb) p *frame > $13 = {root = 0x29f000, parent = 0xadc0de7f3f6c, frames = {next = > 0x50e40de, prev = 0x307f3f88}, local = 0xac00, > this = 0x29f000, ret = 0xadc0de7f3f6c, ref_count = 222, lock = > 84819968, cookie = 0x307f3f88, complete = _gf_false, op = 44032, begin > = { > tv_sec = 11804356835803136, tv_usec = -5926493018029867156}, end = > {tv_sec = 364298988607766750, tv_usec = 52776566472584}, > wind_from = 0xac00 , wind_to > = 0x29f000 , > unwind_from = 0xff7f3f6cbounds>, unwind_to = 0xbounds>} > (gdb) p this->name > $14 = 0x7f3f8803b640 "patchy-tier-dht" (gdb) up #6 0x7f3f8fb5ea98 in dht_lookup_dir_cbk (frame=0x7f3f6c0064bc, cookie=0x7f3f6c0084dc, this=0x7f3f8803aa80, op_ret=-1, op_errno=2, inode=0x7f3f6c0013ec, stbuf=0x7f3f94bc2860, xattr=0x0, postparent=0x7f3f94bc27f0) at /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-common.c:714 714 DHT_STACK_UNWIND (lookup, frame, local->op_ret, local->op_errno, (gdb) p this->name $22 = 0x7f3f88039510 "patchy-hot-dht" (gdb) p frame->parent $23 = (call_frame_t *) 0xadc0de7f3f6c (gdb) p *frame->parent Cannot access memory at address 0xadc0de7f3f6c (gdb) p *frame $24 = {root = 0x64a000, parent = 0xadc0de7f3f6c, frames = {next = 0x50e40de, prev = 0x307f3f88}, local = 0xac00, this = 0x64a000, ret = 0xadc0de7f3f6c, ref_count = 222, lock = 84819968, cookie = 0x307f3f88, complete = _gf_false, op = 44032, begin = { tv_sec = 28323419531509760, tv_usec = -5926493018029867156}, end = {tv_sec = 364298988607766750, tv_usec = 52776566472584}, wind_from = 0xac00 , wind_to = 0x64a000 , unwind_from = 0xff7f3f6c , unwind_to = 0x } (gdb) p this->name $25 = 0x7f3f88039510 "patchy-hot-dht" (gdb) up #7 0x7f3f8fddb23a in client3_3_lookup_cbk (req=0x7f3f6c0090ac, iov=0x7f3f6c0090ec, count=1, myframe=0x7f3f6c0084dc) at /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/protocol/client/src/client-rpc-fops.c:3028 3028 /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/protocol/client/src/client-rpc-fops.c: No such file or directory. (gdb) p frame $26 = (call_frame_t *) 0x7f3f6c0084dc (gdb) p *frame $27 = {root = 0x84c000, parent = 0xadc0de7f3f6c, frames = {next = 0x54d20de, prev = 0x307f3f88}, local = 0xac00, this = 0x84c000, ret = 0xadc0de7f3f6c, ref_count = 222, lock = 88940544, cookie = 0x307f3f88, complete = _gf_false, op = 44032, begin = { tv_sec = 37365803158339584, tv_usec = -5926493018029867156}, end = {tv_sec = 381996727768449246, tv_usec = 52776566472584}, wind_from = 0xac00 , wind_to = 0x84c000 , unwind_from = 0xff7f3f6c , unwind_to = 0x } (gdb) p this->name $28 = 0x7f3f880202d0 "patchy-client-5" Seems like entire stack is corrupted > No, I am not aware of this crash. > > - Original Message - > > From: "Pranith Kumar Karampuri" > > To: "Gluster Devel" , "Raghavendra Gowdappa" > > , "Nithya Balachandran" > > > > Sent: Wednesday, February 10, 2016 5:51:32 AM > > Subject: crash in dht in > > https://build.gluster.org/job/rackspace-regression-2GB-triggered/18134/consoleFull > > > > hi, > > I see the following crash. Is this a known issue? > > (gdb) bt > > #0 0x7f3f8c339fb4 in dht_selfheal_dir_setattr > > (frame=0x7f3f6c002a0c, loc=0x7f3f6c000944, stbuf=0x7f3f6c0009d4, > > valid=16777215, > > layout=0x7f3f6c004140) at > > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-selfheal.c:1087 > > #1 0x7f3f8c33a4f8 in dht_selfheal_dir_mkdir_cbk > > (frame=0x7f3f6c002a0c, cookie=0x7f3f9004201c, this=0x7f3f8803bad0, > > op_ret=-1, op_errno=5, > > inode=0x0, stbuf=0x0, preparent=0x0, postparent=0x0, xdata=0x0) > > at > > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-selfheal.c:1147 > > #2 0x7
Re: [Gluster-devel] changelog bug
btw... he was also having another crash in changelog_rollover: https://gist.githubusercontent.com/CyrilPeponnet/11954cbca725d4b8da7a/raw/2168169f7b208d8ee6193c4a444639505efb634b/gistfile1.txt It would be a pretty huge coincidence if these were all unique causes, wouldn't it? On 02/09/2016 08:27 PM, Kotresh Hiremath Ravishankar wrote: Hi, This crash can't be same as BZ 1221629. The crash in the BZ 1221629 is with the rpc introduced in changelog in 3.7 along with bitrot. Could you share the crash dump to analyse ? Thanks and Regards, Kotresh H R - Original Message - From: "Vijay Bellur" To: "Kotresh Hiremath Ravishankar" , "Manikandan Selvaganesh" Cc: gluster-devel@gluster.org, "cyril peponnet" Sent: Tuesday, February 9, 2016 10:26:57 PM Subject: Re: [Gluster-devel] changelog bug On 02/08/2016 01:14 AM, Kotresh Hiremath Ravishankar wrote: Hi, This bug is already tracked BZ 1221629 I will start working on this and will update once it is fixed. Cyril (in CC) also reported a similar crash with changelog in 3.6.5: https://gist.github.com/CyrilPeponnet/b67b360f186f31d34d8f The crash seems to be consistently reproducible in Cyril's setup. Can we address this soon? Thanks, Vijay ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] crash in dht in https://build.gluster.org/job/rackspace-regression-2GB-triggered/18134/consoleFull
Seems like a double unwind (or frame corruption in tier) (gdb) down #5 0x7f3f8c34c0ac in dht_lookup_dir_cbk (frame=0x7f3f6c002a0c, cookie=0x7f3f6c0064bc, this=0x7f3f8803bad0, op_ret=-1, op_errno=2, inode=0x0, stbuf=0x7f3f6c0069d4, xattr=0x0, postparent=0x7f3f6c006c04) at /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-common.c:724 724 ret = dht_selfheal_directory (frame, dht_lookup_selfheal_cbk, (gdb) p *frame $13 = {root = 0x29f000, parent = 0xadc0de7f3f6c, frames = {next = 0x50e40de, prev = 0x307f3f88}, local = 0xac00, this = 0x29f000, ret = 0xadc0de7f3f6c, ref_count = 222, lock = 84819968, cookie = 0x307f3f88, complete = _gf_false, op = 44032, begin = { tv_sec = 11804356835803136, tv_usec = -5926493018029867156}, end = {tv_sec = 364298988607766750, tv_usec = 52776566472584}, wind_from = 0xac00 , wind_to = 0x29f000 , unwind_from = 0xff7f3f6c , unwind_to = 0x } (gdb) p this->name $14 = 0x7f3f8803b640 "patchy-tier-dht" No, I am not aware of this crash. - Original Message - > From: "Pranith Kumar Karampuri" > To: "Gluster Devel" , "Raghavendra Gowdappa" > , "Nithya Balachandran" > > Sent: Wednesday, February 10, 2016 5:51:32 AM > Subject: crash in dht in > https://build.gluster.org/job/rackspace-regression-2GB-triggered/18134/consoleFull > > hi, > I see the following crash. Is this a known issue? > (gdb) bt > #0 0x7f3f8c339fb4 in dht_selfheal_dir_setattr > (frame=0x7f3f6c002a0c, loc=0x7f3f6c000944, stbuf=0x7f3f6c0009d4, > valid=16777215, > layout=0x7f3f6c004140) at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-selfheal.c:1087 > #1 0x7f3f8c33a4f8 in dht_selfheal_dir_mkdir_cbk > (frame=0x7f3f6c002a0c, cookie=0x7f3f9004201c, this=0x7f3f8803bad0, > op_ret=-1, op_errno=5, > inode=0x0, stbuf=0x0, preparent=0x0, postparent=0x0, xdata=0x0) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-selfheal.c:1147 > #2 0x7f3f8fb81f31 in dht_mkdir (frame=0x7f3f9004201c, > this=0x7f3f8803aa80, loc=0x7f3f6c000944, mode=16877, umask=0, > params=0x7f3f9003594c) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-common.c:6710 > #3 0x7f3f8c33ad1f in dht_selfheal_dir_mkdir (frame=0x7f3f6c002a0c, > loc=0x7f3f6c000944, layout=0x7f3f6c004140, force=0) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-selfheal.c:1256 > #4 0x7f3f8c33c4c8 in dht_selfheal_directory (frame=0x7f3f6c002a0c, > dir_cbk=0x7f3f8c349588 , loc=0x7f3f6c000944, > layout=0x7f3f6c004140) at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-selfheal.c:1837 > #5 0x7f3f8c34c0ac in dht_lookup_dir_cbk (frame=0x7f3f6c002a0c, > cookie=0x7f3f6c0064bc, this=0x7f3f8803bad0, op_ret=-1, op_errno=2, > inode=0x0, > stbuf=0x7f3f6c0069d4, xattr=0x0, postparent=0x7f3f6c006c04) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-common.c:724 > #6 0x7f3f8fb5ea98 in dht_lookup_dir_cbk (frame=0x7f3f6c0064bc, > cookie=0x7f3f6c0084dc, this=0x7f3f8803aa80, op_ret=-1, op_errno=2, > inode=0x7f3f6c0013ec, stbuf=0x7f3f94bc2860, xattr=0x0, > postparent=0x7f3f94bc27f0) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-common.c:714 > #7 0x7f3f8fddb23a in client3_3_lookup_cbk (req=0x7f3f6c0090ac, > iov=0x7f3f6c0090ec, count=1, myframe=0x7f3f6c0084dc) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/protocol/client/src/client-rpc-fops.c:3028 > #8 0x7f3fa0c2d42a in rpc_clnt_handle_reply (clnt=0x7f3f8806cdb0, > pollin=0x7f3f90041c90) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-clnt.c:759 > #9 0x7f3fa0c2d8c8 in rpc_clnt_notify (trans=0x7f3f8806d240, > mydata=0x7f3f8806cde0, event=RPC_TRANSPORT_MSG_RECEIVED, > data=0x7f3f90041c90) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-clnt.c:900 > #10 0x7f3fa0c29b5a in rpc_transport_notify (this=0x7f3f8806d240, > event=RPC_TRANSPORT_MSG_RECEIVED, data=0x7f3f90041c90) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-transport.c:541 > #11 0x7f3f961eadcb in socket_event_poll_in (this=0x7f3f8806d240) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-transport/socket/src/socket.c:2231 > #12 0x7f3f961eb321 in socket_event_handler (fd=18, idx=12, > data=0x7f3f8806d240, poll_in=1, poll_out=0, poll_err=0) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-transport/socket/src/socket.c:2344 >
Re: [Gluster-devel] Gluster test infrastructure and current challenges
On Tue, Feb 9, 2016 at 9:12 PM, Amye Scavarda wrote: > > > On Tue, Feb 9, 2016 at 3:41 PM, Raghavendra Talur > wrote: > >> Hey folks, >> >> Gluster test infrastructure needs a bit of attention. We have committed a >> lot of code in the last couple of years, but we have not scaled our >> infrastructure at the same rate. Lately, the signs have become alarming and >> it calls for attention. >> >> We have scheduled a hangout[1] to educate new Devs about >> 1. Current Gluster test framework - TAP, prove, include.rc. >> 2. Challenges >> a. Regression takes a lot of time to run >> b. We have so many non-deterministic tests >> c. Identifying cause for test failure takes time >> d. NetBSD debugging knowledge in the community is scarce. >> e. DISTAF: Multinode testing is not integrated with regression yet. >> f. More types of tests to have: upgrade, performance, integration. >> g. unit tests >> >> Note that agenda is mainly to showcase the current challenges through a >> medium more informative than email/irc. It is not to discuss possible >> solutions; that should be done over email on the gluster-devel mailing list. >> >> Hence this hangout is scheduled keeping in view convenience of devs from >> IST or around. We will have the session recorded for everyone to view at >> their own convenience. Also we could do the same session for different time >> zones. >> > > This is for tomorrow, Feb 10th? Or am I misreading this event? > With this short of notice, will we be able to get people to attend? > > Thanks! > -amye > > Originally this was supposed to be a record and upload to youtube kind of presentation. We changed it to hangout thinking it could help being a realtime broadcast. Considering it now, I feel we add no value to the event making it a hangout if it is just a demo. We need a hangout/irc meeting for discussion. I will cancel the hangout for now and upload a demo somewhere with the agenda mentioned above.(It would just explain current state) We can discuss in community meeting today for a better date and time to hold a meeting to discuss solutions for challenges that we mention in the demo. Thanks for the input Amye. NOTE: The hangout has been cancelled. We will send invite for the rescheduled session later. >> >> [1] https://plus.google.com/events/c25hcj9llrvvp3dcidqk26sjrhs >> >> Thanks, >> Raghavendra Talur >> >> ___ >> Gluster-devel mailing list >> Gluster-devel@gluster.org >> http://www.gluster.org/mailman/listinfo/gluster-devel >> > > > > -- > Amye Scavarda | a...@redhat.com | Gluster Community Lead > ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] changelog bug
Hi, This crash can't be same as BZ 1221629. The crash in the BZ 1221629 is with the rpc introduced in changelog in 3.7 along with bitrot. Could you share the crash dump to analyse ? Thanks and Regards, Kotresh H R - Original Message - > From: "Vijay Bellur" > To: "Kotresh Hiremath Ravishankar" , "Manikandan > Selvaganesh" > Cc: gluster-devel@gluster.org, "cyril peponnet" > > Sent: Tuesday, February 9, 2016 10:26:57 PM > Subject: Re: [Gluster-devel] changelog bug > > On 02/08/2016 01:14 AM, Kotresh Hiremath Ravishankar wrote: > > Hi, > > > > This bug is already tracked BZ 1221629 > > I will start working on this and will update once it is fixed. > > > > Cyril (in CC) also reported a similar crash with changelog in 3.6.5: > > https://gist.github.com/CyrilPeponnet/b67b360f186f31d34d8f > > The crash seems to be consistently reproducible in Cyril's setup. Can we > address this soon? > > Thanks, > Vijay > > ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
[Gluster-devel] crash in dht in https://build.gluster.org/job/rackspace-regression-2GB-triggered/18134/consoleFull
hi, I see the following crash. Is this a known issue? (gdb) bt #0 0x7f3f8c339fb4 in dht_selfheal_dir_setattr (frame=0x7f3f6c002a0c, loc=0x7f3f6c000944, stbuf=0x7f3f6c0009d4, valid=16777215, layout=0x7f3f6c004140) at /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-selfheal.c:1087 #1 0x7f3f8c33a4f8 in dht_selfheal_dir_mkdir_cbk (frame=0x7f3f6c002a0c, cookie=0x7f3f9004201c, this=0x7f3f8803bad0, op_ret=-1, op_errno=5, inode=0x0, stbuf=0x0, preparent=0x0, postparent=0x0, xdata=0x0) at /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-selfheal.c:1147 #2 0x7f3f8fb81f31 in dht_mkdir (frame=0x7f3f9004201c, this=0x7f3f8803aa80, loc=0x7f3f6c000944, mode=16877, umask=0, params=0x7f3f9003594c) at /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-common.c:6710 #3 0x7f3f8c33ad1f in dht_selfheal_dir_mkdir (frame=0x7f3f6c002a0c, loc=0x7f3f6c000944, layout=0x7f3f6c004140, force=0) at /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-selfheal.c:1256 #4 0x7f3f8c33c4c8 in dht_selfheal_directory (frame=0x7f3f6c002a0c, dir_cbk=0x7f3f8c349588 , loc=0x7f3f6c000944, layout=0x7f3f6c004140) at /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-selfheal.c:1837 #5 0x7f3f8c34c0ac in dht_lookup_dir_cbk (frame=0x7f3f6c002a0c, cookie=0x7f3f6c0064bc, this=0x7f3f8803bad0, op_ret=-1, op_errno=2, inode=0x0, stbuf=0x7f3f6c0069d4, xattr=0x0, postparent=0x7f3f6c006c04) at /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-common.c:724 #6 0x7f3f8fb5ea98 in dht_lookup_dir_cbk (frame=0x7f3f6c0064bc, cookie=0x7f3f6c0084dc, this=0x7f3f8803aa80, op_ret=-1, op_errno=2, inode=0x7f3f6c0013ec, stbuf=0x7f3f94bc2860, xattr=0x0, postparent=0x7f3f94bc27f0) at /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/cluster/dht/src/dht-common.c:714 #7 0x7f3f8fddb23a in client3_3_lookup_cbk (req=0x7f3f6c0090ac, iov=0x7f3f6c0090ec, count=1, myframe=0x7f3f6c0084dc) at /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/xlators/protocol/client/src/client-rpc-fops.c:3028 #8 0x7f3fa0c2d42a in rpc_clnt_handle_reply (clnt=0x7f3f8806cdb0, pollin=0x7f3f90041c90) at /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-clnt.c:759 #9 0x7f3fa0c2d8c8 in rpc_clnt_notify (trans=0x7f3f8806d240, mydata=0x7f3f8806cde0, event=RPC_TRANSPORT_MSG_RECEIVED, data=0x7f3f90041c90) at /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-clnt.c:900 #10 0x7f3fa0c29b5a in rpc_transport_notify (this=0x7f3f8806d240, event=RPC_TRANSPORT_MSG_RECEIVED, data=0x7f3f90041c90) at /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-transport.c:541 #11 0x7f3f961eadcb in socket_event_poll_in (this=0x7f3f8806d240) at /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-transport/socket/src/socket.c:2231 #12 0x7f3f961eb321 in socket_event_handler (fd=18, idx=12, data=0x7f3f8806d240, poll_in=1, poll_out=0, poll_err=0) at /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-transport/socket/src/socket.c:2344 #13 0x7f3fa0ec61a8 in event_dispatch_epoll_handler (event_pool=0x121fce0, event=0x7f3f94bc2e70) at /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/libglusterfs/src/event-epoll.c:571 #14 0x7f3fa0ec6596 in event_dispatch_epoll_worker (data=0x125e1a0) at /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/libglusterfs/src/event-epoll.c:674 #15 0x7f3fa0144a51 in start_thread () from ./lib64/libpthread.so.0 #16 0x7f3f9faae93d in clone () from ./lib64/libc.so.6 Pranith ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] changelog bug
On 02/08/2016 01:14 AM, Kotresh Hiremath Ravishankar wrote: Hi, This bug is already tracked BZ 1221629 I will start working on this and will update once it is fixed. Cyril (in CC) also reported a similar crash with changelog in 3.6.5: https://gist.github.com/CyrilPeponnet/b67b360f186f31d34d8f The crash seems to be consistently reproducible in Cyril's setup. Can we address this soon? Thanks, Vijay ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] Gluster test infrastructure and current challenges
On Tue, Feb 9, 2016 at 3:41 PM, Raghavendra Talur wrote: > Hey folks, > > Gluster test infrastructure needs a bit of attention. We have committed a > lot of code in the last couple of years, but we have not scaled our > infrastructure at the same rate. Lately, the signs have become alarming and > it calls for attention. > > We have scheduled a hangout[1] to educate new Devs about > 1. Current Gluster test framework - TAP, prove, include.rc. > 2. Challenges > a. Regression takes a lot of time to run > b. We have so many non-deterministic tests > c. Identifying cause for test failure takes time > d. NetBSD debugging knowledge in the community is scarce. > e. DISTAF: Multinode testing is not integrated with regression yet. > f. More types of tests to have: upgrade, performance, integration. > g. unit tests > > Note that agenda is mainly to showcase the current challenges through a > medium more informative than email/irc. It is not to discuss possible > solutions; that should be done over email on the gluster-devel mailing list. > > Hence this hangout is scheduled keeping in view convenience of devs from > IST or around. We will have the session recorded for everyone to view at > their own convenience. Also we could do the same session for different time > zones. > This is for tomorrow, Feb 10th? Or am I misreading this event? With this short of notice, will we be able to get people to attend? Thanks! -amye > > > [1] https://plus.google.com/events/c25hcj9llrvvp3dcidqk26sjrhs > > Thanks, > Raghavendra Talur > > ___ > Gluster-devel mailing list > Gluster-devel@gluster.org > http://www.gluster.org/mailman/listinfo/gluster-devel > -- Amye Scavarda | a...@redhat.com | Gluster Community Lead ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
[Gluster-devel] Gluster test infrastructure and current challenges
Hey folks, Gluster test infrastructure needs a bit of attention. We have committed a lot of code in the last couple of years, but we have not scaled our infrastructure at the same rate. Lately, the signs have become alarming and it calls for attention. We have scheduled a hangout[1] to educate new Devs about 1. Current Gluster test framework - TAP, prove, include.rc. 2. Challenges a. Regression takes a lot of time to run b. We have so many non-deterministic tests c. Identifying cause for test failure takes time d. NetBSD debugging knowledge in the community is scarce. e. DISTAF: Multinode testing is not integrated with regression yet. f. More types of tests to have: upgrade, performance, integration. g. unit tests Note that agenda is mainly to showcase the current challenges through a medium more informative than email/irc. It is not to discuss possible solutions; that should be done over email on the gluster-devel mailing list. Hence this hangout is scheduled keeping in view convenience of devs from IST or around. We will have the session recorded for everyone to view at their own convenience. Also we could do the same session for different time zones. [1] https://plus.google.com/events/c25hcj9llrvvp3dcidqk26sjrhs Thanks, Raghavendra Talur ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
[Gluster-devel] Minutes of today's Gluster Community Bug Triage Meeting
Hi all, Thanks everyone for attending the Gluster Community Bug Triage meeting today and here are the minutes of the meeting: Meeting summary --- * agenda: https://public.pad.fsfe.org/p/gluster-bug-triage (Manikandan, 12:00:41) * Roll Call (Manikandan, 12:00:48) * kkeithley_ will come up with a proposal to reduce the number of bugs against "mainline" in NEW state (Manikandan, 12:05:23) * ACTION: kkeithley_ will come up with a proposal to reduce the number of bugs against "mainline" in NEW state (Manikandan, 12:06:15) * msvbhat and ndevos need to think about and decide how to provide/use debug builds (Manikandan, 12:06:35) * ACTION: msvbhat and ndevos need to think about and decide how to provide/use debug build (Manikandan, 12:07:34) * Group Triage (Manikandan, 12:08:38) * LINK: http://gluster.readthedocs.org/en/latest/Contributors-Guide/Bug-Triage/ (Manikandan, 12:09:07) * Open Floor (Manikandan, 12:26:23) Meeting ended at 12:29:30 UTC. Action Items * kkeithley_ will come up with a proposal to reduce the number of bugs against "mainline" in NEW state * msvbhat and ndevos need to think about and decide how to provide/use debug build. * kkeithley_ will come up with a proposal to reduce the number of bugs against "mainline" in NEW state * hagarth start/sync email on regular (nightly) automated tests * msvbhat will look into using nightly builds for automated testing, and will report issues/success to the mailinglist * msvbhat will look into lalatenduM's automated Coverity setup in Jenkins whichneed assistance from an admin with more permissions * msvbhat and ndevos need to think about and decide how to provide/use debug builds * msvbhat provide a simple step/walk-through on how to provide testcases for the nightly rpm tests * ndevos to propose some test-cases for minimal libgfapi test * Manikandan and Nandaja will keep updating on the bug automation workflow. People Present (lines said) --- * Manikandan (43) * ndevos (6) * hgowtham (5) * skoduri (4) * jiffin (3) * zodbot (3) * Saravanakmr (3) See you all next week :-) -- Thanks & Regards, Manikandan Selvaganesh. ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
[Gluster-devel] REMINDER: Gluster Community Bug Triage meeting at 12:00 UTC (~in 60 minutes)
Hi all, This meeting is scheduled for anyone that is interested in learning more about, or assisting with the Bug Triage. Meeting details: - location: #gluster-meeting on Freenode IRC (https://webchat.freenode.net/?channels=gluster-meeting ) - date: every Tuesday - time: 12:00 UTC (in your terminal, run: date -d "12:00 UTC") - agenda: https://public.pad.fsfe.org/p/gluster-bug-triage Currently the following items are listed: * Roll Call * Status of last weeks action items * Group Triage * Open Floor The last two topics have space for additions. If you have a suitable bug or topic to discuss, please add it to the agenda. Appreciate your participation. Thank you :-) -- Regards, Manikandan Selvaganesh. ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] Cores on NetBSD of brick https://build.gluster.org/job/rackspace-netbsd7-regression-triggered/14100/consoleFull
On 02/09/2016 04:13 PM, Emmanuel Dreyfus wrote: On Tue, Feb 09, 2016 at 11:56:37AM +0530, Pranith Kumar Karampuri wrote: I think the regression run is not giving that link anymore when the crash happens? Could you please add that also as a link in regression run? Ther was the path of the archive, I changed it for a http:// link Oops, sorry, it is there. Pranith ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel
Re: [Gluster-devel] Cores on NetBSD of brick https://build.gluster.org/job/rackspace-netbsd7-regression-triggered/14100/consoleFull
On Tue, Feb 09, 2016 at 11:56:37AM +0530, Pranith Kumar Karampuri wrote: > I think the regression run is not giving that link anymore when the crash > happens? Could you please add that also as a link in regression run? Ther was the path of the archive, I changed it for a http:// link -- Emmanuel Dreyfus m...@netbsd.org ___ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel