On 2014-08-01 08:56, Pranith Kumar Karampuri wrote: > > On 08/01/2014 12:09 PM, Anders Blomdell wrote: >> On 2014-08-01 02:02, Harshavardhana wrote:> On Thu, Jul 31, 2014 at 11:31 >> AM, Anders Blomdell >>> <anders.blomd...@control.lth.se> wrote: >>>> During rsync of 350000 files, memory consumption of glusterfs >>>> rose to 12 GB (after approx 14 hours), I take it that this is a >>>> bug I should try to track down? >>>> >>> Does it ever come down? what happens if you repeatedly the same files >>> again? does it OOM? >> Well, it OOM'd my firefox first (that's how good I monitor my experiments >> :-() >> No, memory usage does not come down by itself, ACAICT >> >> >> On 2014-08-01 02:12, Raghavendra Gowdappa wrote:> Anders, >>> Mostly its a case of memory leak. It would be helpful if you can file a bug >>> on this. Following information would be useful to fix the issue: >>> >>> 1. valgrind reports (if possible). >>> a. To start brick and nfs processes with valgrind you can use following >>> cmdline when starting glusterd. >>> # glusterd --xlator-option *.run-with-valgrind=yes >>> >>> In this case all the valgrind logs can be found in standard glusterfs >>> log directory. >>> >>> b. For client you can start glusterfs just like any other process in >>> valgrind. Since glusterfs is daemonized, while running with valgrind we >>> need to prevent it by running it in foreground. We can use -N option to do >>> that >>> # valgrind --leak-check=full --log-file=<path-to-valgrind-log> >>> glusterfs --volfile-id=xyz --volfile-server=abc -N /mnt/glfs >>> >>> 2. Once you observe a considerable leak in memory, please get a statedump >>> of glusterfs >>> >>> # gluster volume statedump <volname> >>> >>> and attach the reports in the bug. >> Since it looks like Pranith has a clue, I'll leave it for a few weeks (other >> pressing duties). >> >> On 2014-08-01 03:24, Pranith Kumar Karampuri wrote: >>> Yes, even I saw the following leaks, when I tested it a week back. These >>> were the leaks: >>> You should probably take a statedump and see what datatypes are leaking. >>> >>> root@localhost - /usr/local/var/run/gluster >>> 14:10:26 ? awk -f /home/pk1/mem-leaks.awk glusterdump.22412.dump.1406174043 >>> [mount/fuse.fuse - usage-type gf_common_mt_char memusage] >>> size=341240 >>> num_allocs=23602 >>> max_size=347987 >>> max_num_allocs=23604 >>> total_allocs=653194 >>> ... >> I'll revisit this in a few weeks, >> >> Harshavardhana, Raghavendra, Pranith (and all others), >> >> Gluster is one of the most responsive Open Source project I >> have participated in this far, I'm very happy with all support, >> help and encouragement I have got this far. Even though my initial >> tests weren't fully satisfactory, you are the main reason for >> my perseverance :-) > Yay! good :-). Do you have any suggestions where we need to improve as > a community that would make it easier for new contributors? http://review.gluster.org/#/c/8181/ (will hopefully come around and review that, real soon now...)
Otherwise, no. Will recommend gluster as an eminent crash-course in git, gerrit and continous integration. Keep up the good work. /Anders -- Anders Blomdell Email: anders.blomd...@control.lth.se Department of Automatic Control Lund University Phone: +46 46 222 4625 P.O. Box 118 Fax: +46 46 138118 SE-221 00 Lund, Sweden _______________________________________________ Gluster-devel mailing list Gluster-devel@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-devel