Re: [Gluster-devel] Weekly Community Meeting - 20170104

2017-01-04 Thread Kaushal M
Good start to 2017. We had an active meeting this time.

The meeting minutes and weekly updates are available below, as well as
the links to the logs. The meeting agenda and updates have been
archived at 
https://github.com/gluster/glusterfs/wiki/Community-Meeting-2017-01-04
.
Next weeks meeting will be hosted by me, at the same time and place.
1200UTC in #gluster-meeting on Freenode. The agenda is now open for
topics and updates at https://bit.ly/gluster-community-meetings .

See you all next week.

~kaushal

Minutes: 
https://meetbot.fedoraproject.org/gluster-meeting/2017-01-04/gluster_community_meeting_20170104.2017-01-04-12.00.html
Minutes (text):
https://meetbot.fedoraproject.org/gluster-meeting/2017-01-04/gluster_community_meeting_20170104.2017-01-04-12.00.txt
Log: 
https://meetbot.fedoraproject.org/gluster-meeting/2017-01-04/gluster_community_meeting_20170104.2017-01-04-12.00.log.html

## Updates

> NOTE : Updates will not be discussed during meetings. Any important or 
> noteworthy update will be announced at the end of the meeting

### Action Items from last week


- Discuss participation in the meetings in January.
- Carryover to January

### Releases

 GlusterFS 4.0

- Tracker bug :
https://bugzilla.redhat.com/showdependencytree.cgi?id=glusterfs-4.0
- Roadmap : https://www.gluster.org/community/roadmap/4.0/
- Updates:
- GD2
- New release GlusterD2 v4.0dev-4
- More details on the work done at
https://www.gluster.org/pipermail/gluster-devel/2017-January/051805.html

 GlusterFS 3.10

- Maintainers : shyam, kkeithley, rtalur
- Next release : 3.10.0
- Target date: February 14, 2017
- Release tracker : https://github.com/gluster/glusterfs/milestone/1
- Updates:
  - Feature list frozen, link as above
  - Branching date: 17th Jan, 2017 (~4 weeks prior to the release date
of 14th Feb, 2017)
  - Feature readiness checkpoint will be done around 3rd/4th Jan, 2017
  - Call out: Feature specs need reviews and closure (will be sending
a mail regarding the same)
  - Reference mail:
http://www.gluster.org/pipermail/gluster-devel/2016-December/051674.html

 GlusterFS 3.9

- Maintainers : pranithk, aravindavk, dblack
- Current release : 3.9.0
- Next release : 3.9.1
  - Release date : 20 January 2017
- Tracker bug :
https://bugzilla.redhat.com/show_bug.cgi?id=glusterfs-3.9.1 (doesn't
exist)
- Open bugs : 
https://bugzilla.redhat.com/showdependencytree.cgi?maxdepth=2=glusterfs-3.9.0_resolved=1
- Roadmap : https://www.gluster.org/community/roadmap/3.9/
- Updates:
  - _None_

 GlusterFS 3.8

- Maintainers : ndevos, jiffin
- Current release : 3.8.7
- Next release : 3.8.8
  - Release date : 10 January 2017
- Tracker bug : https://bugzilla.redhat.com/show_bug.cgi?id=glusterfs-3.8.8
- Open bugs : 
https://bugzilla.redhat.com/showdependencytree.cgi?maxdepth=2=glusterfs-3.8.8_resolved=1
- Updates:
  - _None_

 GlusterFS 3.7

- Maintainers : kshlm, samikshan
- Current release : 3.7.18
- Next release : 3.7.19
  - Release date : 30 December 2016
- Tracker bug : https://bugzilla.redhat.com/show_bug.cgi?id=glusterfs-3.7.19
- Open bugs : 
https://bugzilla.redhat.com/showdependencytree.cgi?maxdepth=2=glusterfs-3.7.19_resolved=1
- Updates:
  - 3.7.18 was finally tagged a week late
  - Expect release/announcement later this week
  - Announcement done on 2016-12-13
  - 
https://www.gluster.org/pipermail/gluster-users/2016-December/029427.html
  - 3.7.19 should be tagged later this week (week of 20170101)

### Related projects and efforts

 Community Infra

- The Gerrit OS upgrade didn't go through during the holidays. We'll
be schduling it this month at an appropriate weekend.
- fstat also [shows branch
names](http://fstat.gluster.org/weeks/4/failure/82) now so you know if
a failure happens only in one specific branch. It only does this for
new jobs, not for old ones. (If there is demand, I'll add it for old
ones)
- When a NetBSD jobs is aborted, the machine will now be automatically
restarted.
- We've addded an additional machine for netbsd smoke since the queue
was getting quite long with just one machine.

 Samba

- _None_

 Ganesha

- _None_

 Containers

- _None_

 Testing

- _None_

 Others

- aravindavk, Updates on Geo-replication
- https://www.gluster.org/pipermail/gluster-devel/2016-December/051636.html
- Top 5 regressions in
[December](https://www.gluster.org/pipermail/gluster-devel/2016-December/051792.html)


Meeting summary
---
* Rollcall  (kshlm, 12:00:38)

* STM and backports  (kshlm, 12:06:18)
  * ACTION: Need to find out when 3.9.1 is happening  (kshlm, 12:17:51)

* A common location for testing-tools  (kshlm, 12:18:20)
  * ACTION: shyam will file a bug to get arequal included in glusterfs
packages  (kshlm, 12:42:46)

* Developer workflow problems  (kshlm, 12:43:16)

Meeting ended at 13:10:03 UTC.




Action Items

* Need to find out when 3.9.1 is happening
* shyam will file a bug to get arequal included in 

Re: [Gluster-devel] Assertion failed: lru_inode_ctx->block_num > 0

2017-01-04 Thread qingwei wei
Hi Krutika,

Happy new year to you!

Regarding this issue, do you have any new update?

Cw

On Fri, Dec 23, 2016 at 1:05 PM, Krutika Dhananjay  wrote:
> Perfect. That's what I needed to know. Thanks! :)
>
> -Krutika
>
> On Fri, Dec 23, 2016 at 7:15 AM, qingwei wei  wrote:
>>
>> Hi Krutika,
>>
>> Is this information you need?
>>
>> [root@localhost /]# gdb /usr/local/sbin/glusterfs core.14019
>> GNU gdb (GDB) Red Hat Enterprise Linux 7.6.1-80.el7
>> Copyright (C) 2013 Free Software Foundation, Inc.
>> License GPLv3+: GNU GPL version 3 or later
>> 
>> This is free software: you are free to change and redistribute it.
>> There is NO WARRANTY, to the extent permitted by law.  Type "show copying"
>> and "show warranty" for details.
>> This GDB was configured as "x86_64-redhat-linux-gnu".
>> For bug reporting instructions, please see:
>> ...
>> Reading symbols from /usr/local/sbin/glusterfsd...done.
>>
>> warning: core file may not match specified executable file.
>> [New LWP 14025]
>> [New LWP 14019]
>> [New LWP 14020]
>> [New LWP 14029]
>> [New LWP 14021]
>> [New LWP 14022]
>> [New LWP 14023]
>> [New LWP 14024]
>> [New LWP 14028]
>> [Thread debugging using libthread_db enabled]
>> Using host libthread_db library "/lib64/libthread_db.so.1".
>> Core was generated by `/usr/local/sbin/glusterfs
>> --volfile-server=10.217.242.32 --volfile-id=/testSF1'.
>> Program terminated with signal 11, Segmentation fault.
>> #0  list_del_init (old=0x7f5d6c003d10) at
>> ../../../../libglusterfs/src/list.h:87
>> 87old->prev->next = old->next;
>> (gdb) select-frame 3
>> (gdb) print local->fop
>> $1 = GF_FOP_WRITE
>> (gdb)
>>
>> Hopefully is useful for your investigation.
>>
>> Thanks.
>>
>> Cw
>>
>>
>> On Wed, Dec 21, 2016 at 8:13 PM, Krutika Dhananjay 
>> wrote:
>> > Just one more information I need from you.
>> > Assuming you have the coredump, could you attach it to gdb and print
>> > local->fop and tell me what fop it was when the crash happened?
>> > You'll need to switch to frame 3 in gdb to get the value of this
>> > variable.
>> >
>> > -Krutika
>> >
>> > On Wed, Dec 21, 2016 at 5:35 PM, Krutika Dhananjay 
>> > wrote:
>> >>
>> >> Thanks for this. The information seems sufficient at the moment.
>> >> Will get back to you on this if/when I find something.
>> >>
>> >> -Krutika
>> >>
>> >> On Mon, Dec 19, 2016 at 1:44 PM, qingwei wei 
>> >> wrote:
>> >>>
>> >>> Hi Krutika,
>> >>>
>> >>> Sorry for the delay as i am busy with other works. Attached is the
>> >>> tar.gz file with client and server log, the gfid information on the
>> >>> shard folder (please look at test.0.0 file as the log is captured when
>> >>> i run fio on this file.) and also the print statement i put inside the
>> >>> code. Fyi, i did 2 runs this time and only the second run give me
>> >>> problem. Hope this information helps.
>> >>>
>> >>> Regards,
>> >>>
>> >>> Cw
>> >>>
>> >>> On Thu, Dec 15, 2016 at 8:02 PM, Krutika Dhananjay
>> >>> 
>> >>> wrote:
>> >>> > Good that you asked. I'll try but be warned this will involve me
>> >>> > coming
>> >>> > back
>> >>> > to you with lot more questions. :)
>> >>> >
>> >>> > I've been trying this for the past two days (not to mention the fio
>> >>> > run
>> >>> > takes
>> >>> > really long) and so far there has been no crash/assert failure.
>> >>> >
>> >>> > If you already have the core:
>> >>> > in frame 1,
>> >>> > 0. print block_num
>> >>> > 1. get lru_inode_ctx->stat.ia_gfid
>> >>> > 2. convert it to hex
>> >>> > 3. find the gfid in your backend that corresponds to this gfid and
>> >>> > share its
>> >>> > path in your response
>> >>> > 4. print priv->inode_count
>> >>> > 5. and of course lru_inode_ctx->block_num :)
>> >>> > 6. Also attach the complete brick and client logs.
>> >>> >
>> >>> > -Krutika
>> >>> >
>> >>> >
>> >>> > On Thu, Dec 15, 2016 at 3:18 PM, qingwei wei 
>> >>> > wrote:
>> >>> >>
>> >>> >> Hi Krutika,
>> >>> >>
>> >>> >> Do you need anymore information? Do let me know as i can try on my
>> >>> >> test system. Thanks.
>> >>> >>
>> >>> >> Cw
>> >>> >>
>> >>> >> On Tue, Dec 13, 2016 at 12:17 AM, qingwei wei 
>> >>> >> wrote:
>> >>> >> > Hi Krutika,
>> >>> >> >
>> >>> >> > You mean FIO command?
>> >>> >> >
>> >>> >> > Below is how i do the sequential write. This example i am using
>> >>> >> > 400GB
>> >>> >> > file, for the SHARD_MAX_INODE=16, i use 300MB file.
>> >>> >> >
>> >>> >> > fio -group_reporting -ioengine libaio -directory /mnt/testSF-HDD1
>> >>> >> > -fallocate none -direct 1 -filesize 400g -nrfiles 1 -openfiles 1
>> >>> >> > -bs
>> >>> >> > 256k -numjobs 1 -iodepth 2 -name test -rw write
>> >>> >> >
>> >>> >> > And after FIO complete the above workload, i do the random write
>> >>> >> >
>> >>> >> > fio -group_reporting -ioengine libaio -directory